var/home/core/zuul-output/0000755000175000017500000000000015134405717014534 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015134417122015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000236615215134416766020302 0ustar corecorerikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD >"mv?_eGbuuțx{w7ݭ7֫'e% oo/q3m^]/o?8.7oW}ʋghewx/mX,ojŻ ^Tb3b#׳:}=p7뼝ca㑔`e0I1Q!&ѱ[/o^{W-{t3_U|6 x)K#/5ΌR"ggóisR)N %emOQ/Ϋ[oa0vs68/Jʢ ܚʂ9ss3+aô٥J}{37FEbп3 FKX1QRQlrTvb)E,s)Wɀ;$#LcdHM%vz_. o~I|3j dF{ "IΩ?PF~J~ ` 17ׅwڋًM)$Fiqw7Gt7L"u 0V9c  ˹dvYļU[ Z.׿-h QZ*U1|t5wKOؾ{mk b2 ܨ;RJK!b>JR*kl|+"N'C_#a7]d]sJg;;>Yp׫,w`ɚ'd$ecwŻ^~7EpQС3DCS[Yʧ?DDS aw߾)VxX帟AB}nyи0stĈCo.:wAZ{sy:7qsWctx{}n-+ZYsI{/.Ra9XcђQ0FK@aEDO2es ׇN# ZF͹b,*YVi+$<QMGhC}^}?BqG!(8l K3T[<~6]90}(*T7siv'=k 9Q2@vN ( R['>v*;o57sp$3ncx!>t®W>]tF-iܪ%GYbaRvHa}dkD̶*';ک|s_}8yj,('GrgTZ'U鋊TqOſ * /Ijo!՟8`"j}zӲ$k3jS|C7;A)͎V.r?t\WU1ojjr<~Tq> `=tJ!aݡ=h6Yݭw}?lѹ`f_" J9w4ts7NG GGG]ҡgc⌝M b/Ζlpah E ur C&`XR JcwB~R2EL9j7e\(Uё$׿atyХ?*t5z\+`/ErVQUxMҔ&ۈt.3;eg_O ξL1KiYLizpV:C5/=v-}҅"o ']쌕|tϓX8nJ*A*%J[T2pI1Je;s_[,Ҩ38_ь ͰM0ImY/MiVJ5&jNgBt90v߁R:~U jځU~oN9xԞ~J|dݤ߯R> kH&Y``:"s ayiBq)u%'4 yܽ yW0 -i̭uJ{KưЖ@+UBj -&JO x@}DS.€>3T0|9ē7$3z^.I< )9qf e%dhy:O40n'c}c1XҸuFiƠIkaIx( +")OtZ l^Z^CQ6tffEmDφǽ{QiOENG{P;sHz"G- >+`قSᔙD'Ad ѭj( ہO r:91v|ɛr|٦/o{C Ӹ!uWȳ)gjw&+uߕt*:͵UMQrN@fYDtEYZb4-UCqK٪L.2teB ˛"ո{Gci`du듎q+;C'16FgVlWaaB)"F,u@30YQg˾_YҊŏ#_f^ TD=VAKNl4Kš4GScѦa0 J ()¾5m'p/\խX\=z,Mw˭x:qu礛WԓL!I xӤ1(5AKRVF2ɌУլ F "vuhc=JS\kkZAY`R"Hr1]%oR[^oI]${&L8<=#0yaKL: JJl r;t#H+B|ɧJiM cm)>H=l}.^\ݧM<lu Y> XH\z:dHElL(uHR0i#q%]!=t_쾋-, vW~* ^g/5n]FhNU˿oۂ6C9C7sn,kje*;iΓA7,Q)-,=1A sK|ۜLɽy]ʸEO<-YEqKzϢ \{>dDLF amKGm+`VLJsC>?5rk{-3Ss`y_C}Q v,{*)ߎ% qƦat:D=uNvdߋ{Ny[$ {ɴ6hOI']dC5`t9:GO: FmlN*:g^;T^B0$B%C6Θ%|5u=kkN2{'FEc* A>{avdt)8|mg定TN7,TEXt+`F P |ɧ<Ғ8_iqE b}$B#fethBE;1"l r  B+R6Qp%;R8P󦟶Ub-L::;Ⱦ7,VW.JE:PgXoΰUv:ΰdɆΰ (ΰ0eTUgXun[g, ׽-t!X򴱞_aM:E.Qg1DllЊE҉L ehJx{̗Uɾ?si&2"C]u$.`mjmƒVe9f6NŐsLu6fe wkىKR%f"6=rw^)'Hz }x>1yFX09'A%bDb0!i(`Z;TyֻΗ|ִ0-6dAC5t[OM91c:VJR9&ksvJ;0ɝ$krogB= FYtЩOte=?>T&O{Ll)HClba1PIFĀ":tu^}.&R*!^pHPQuSVO$.KMb.:DK>WtWǭKv4@Va3"a`R@gbu%_J5Ґ 3?lm$K/$s_. WM]̍"W%`lO2-"ew@E=! I,($F{ձ7*Oy 6EK( EF #31J8mN .TTF9㕴/5~RxCe,&v3,JE- ZF5%Da,Gܠ*qI@qlG6s푻jÝ$ >8ȕ$eZ1j[h0SH,qf<"${/ksBK}xnwDb%M6:K<~̓9*u᛹Q{FЖt~6S#G1(zr6<ߜ!?U\(0EmG4 4c~J~]ps/9܎ms4gZY-07`-Id,9õ԰t+-b[uemNi_󈛥^g+!SKq<>78NBx;c4<ニ)H .Pd^cR^p_G+E--ۥ_F]a|v@|3p%kzh|k*BBRib\J3Yn|뇱[FfP%M:<`pz?]6laz5`ZQs{>3ư_o%oU׆]YLz_s߭AF'is^_&uUm$[[5HI4QCZ5!N&D[uiXk&2Bg&Ս7_/6v_cd쿽d@eU XyX2z>g8:.⺻h()&nO5YE\1t7aSyFxPV19 ĕi%K"IcB j>Pm[E[^oHmmU̸nG pHKZ{{Qo}i¿Xc\]e1e,5`te.5Hhao<[50wMUF􀍠PV?Yg"ź)\3mf|ܔMUiU|Ym! #'ukMmQ9Blm]TO1ba.XW x6ܠ9[v35H;-]Um4mMrW-k#~fؤϋu_j*^Wj^qM `-Pk.@%=X#|ۡb1lKcj$׋bKv[~"N jS4HOkeF3LPyi︅iWk! cAnxu6<7cp?WN $?X3l(?  'Z! ,Z.maO_Bk/m~ޖ(<qRfR"Au\PmLZ"twpuJ` mvf+T!6Ѓjw1ncuwo':o gSPC=]U҅yY9 &K<-na'Xk,P4+`Þ/lX/bjFO.= w ?>ȑ3n߿z,t s5Z/ Clo-` z?a~b mzkC zFȏ>1k*Dls6vP9hS  ehC.3 @6ijvUuBY hBnb[ Fr#D7ćlA!:X lYE>#0JvʈɌ|\u,'Y˲.,;oOwoj-25Hݻ7 li0bSlbw=IsxhRbd+I]Y]JP}@.供SЃ??w w@KvKts[TSa /ZaDžPAEư07>~w3n:U/.P珀Yaٳ5Ʈ]խ4 ~fh.8C>n@T%W?%TbzK-6cb:XeGL`'žeVVޖ~;BLv[n|viPjbMeO?!hEfޮ])4 ?KN1o<]0Bg9lldXuT ʑ!Iu2ʌnB5*<^I^~G;Ja߄bHȌsK+D"̽E/"Icƀ@`<ۃ7|qk" L+Y*Ha)j~pu7ި!:E#s:ic.XC^wT/]n2'>^&pnapckL>2QQWo/ݻ<̍8)r`F!Woc0Xq0 R' eQ&Aѣzvw=e&".awfShWjÅD0JkBh]s9Ą|ק_;%X6Q@d 8&a)a.#ۿD> vfA{$g ăyd) SK?ɧ#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhhdEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc0]rY9'z .(jHI :{HG}HDN`h7@{jnE#[dz;n#y 9D*A$$"^)dVQ.(rO6ӟZw_Ȣaޒu'- ^_,G;U\cAAz7EtlLuoXuA}bT2H_*kIG?S(קjhg 5EF5uKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxtK|v+`VZ3JϧC^|/{ś}r3 >6׳oƄ%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwG%mgo 恤hˍJ_SgskwI\t`ﶘ080ƱQŀllKX@116fqo>NrU Ѣ9*|ãeeH7.z!<7zG4p9tV|̢T`˖E ;;,tTaIUle*$!>*mBA2,gJIn_kSz)JC]?X(OPJS3.}clݨ{e!MB,cB߮4af祋,1/_xq=fBRO0P'֫-kbM6Apw,GO2}MGK'#+սE^dˋf6Y bQEz}eҏnr_ ^O^W zw~Ȳ=sXअy{E|r,8.7uO`c Nc0%Ն R C%_ EV a"҅4 |T!DdǍ- .™5,V:;[g./0 +v䤗dWF >:֓[@ QPltsHtQ$J==O!;*>ohǖVa[|E7e0ϕ9Uyzg%pg/cc6RS`HFLЩ LkJu\!`0);Sak$Vfp~C%YdE6c>1ƕ (0W4Q>@>lWN"^ X5G-nm.8B>NOI[31,j2 Ce |M>8l WIf|\q4|UkC.gr`˱Lϰ} xr.~l-ɩu_Drd31V_ѺUib0/ %IYhq ҕ  O UA!wY~ -`%Űb`\mS38W1`vOF7/.C!Pu&Jm l?Q>}O+D7 P=x@`0ʿ26a>d Bqε^a'NԋsI`Yu.7v$Rt)Ag:ݙyX|HkX cU82IP qgzkX=>׻K߉J%E92' ]qҙ%rXgs+"sc9| ]>T]"JرWBΌ-zJS-~y30G@U#=h7) ^EUB Q:>9W΀çM{?`c`uRljצXr:l`T~IQg\Ѝpgu#QH! ,/3`~eB|C1Yg~ؼ/5I7w9I}qww}U~7뭱ԏ,}e7]ukDn`jSlQ7DžHa/EU^IpYWW兹Q7WyTz|nˇ _qˍ[!;n ^b k[);ng]ȶM_u)O_xV hx h[K2kـ`b duhq[..cS'5YO@˒ӓdcY'HAKq^$8`b $1r Qz?ۧ1ZM/G+qYcYl YhD$kt_TId E$dS:֢̆ ?GЅ'JƖ'ZXO݇'kJՂU086\h%1GK(Yn% ']Q; Gd:!gI-XEmkF}:~0}4t3Qf5xd\hEB-} |q*ȃThLj'sQ %؇Gk`F;Sl\h)5؈x2Ld="KԦ:EVewN ًS9d#$*u>>I#lX9vW !&H2kVyKZt<cm^] bCD6b&>9VE7e4p +{&g߷2KY,`Wf1_ܑMYٚ'`ySc4ΔV`nI+ƳC6;җ2ct"*5S}t)eNqǪP@o`co ˎ<عLۀG\ 7۶+q|YRiĹ zm/bcK3;=,7}RqT vvFI O0]&5uKMf#pDTk6yi*cem:y0W|1u CWL;oG^\ X5.aRߦ[_Vs? Ž^A12JQ̛XL:OEUپOY>WK-uP0\8"M: /P4Qz~j3 .-8NJ|!N9/|a|>lX9T ҇t~T1=UF"t; 8-1I|2L+)WȱL˿ˍ-038D*0-)ZyT13`tTnm|Yhi+lQ&Z!֨řoҒ"HKX 6„=z{Ҍ5+P1;ڇ6UNE@Uo/>8.fgW]kY0Cgcu6/!_Ɩ} ' Ў3)X<seWfSv!ؒRKfs%(1Lhrٵ L.] s?I,HCԢ[b C-lLG+@_$c%* _jR|\:dc5u= A@kUc\ǔz;M>dUN/aFRĦ@x؂ǀ$6%}N^ \mQ!%8j0dUo=rh>*YȴU3Q,̸*E%59sTzɟڮ2kg ۱wEUD3uKrr&"B:p`\E)j<).R&#ÃecE,dp"nPS 44 Q8ZƈKnnJei+^z '3JDbSK;*uБ:hF ѹ @˿ޗ~7g9| hLXULi7.1-Qk%ƩJ4^=ple;u.6vQe UZAl *^Vif]>HUd6ƕ̽=T/se+ϙK$S`hnOcE(Tcr!:8UL | 8 !t Q7jk=nn7J0ܽ0{GGL'_So^ʮL_'s%eU+U+ȳlX6}i@djӃfb -u-w~ r}plK;ֽ=nlmuo[`wdй d:[mS%uTڪ?>={2])|Ը>U{s]^l`+ ja^9c5~nZjA|ЩJs Va[~ۗ#rri# zLdMl?6o AMҪ1Ez&I2Wwߎ|7.sW\zk﯊溺^TW^T\*6eqr/^T77WNZ7F_}-򲺺VWQ77V\_v?9?"Th $LqQjiXMlk1=VzpO֠24hf 1hi D{q:v%̈#v^nBi~MefZF >:/?Ac 1M'I`22؆DT!/j璓P åiw@wgRCsT~$U>ceއE)BI>UljO|Ty$ŋrwOtZ7$ "i 8U 7bSem'k?I+/޵$"弄ټ0316N|;`I6%S$Ëlϯ?U$%ˎULXb'ŮuR4I=Jb"E2[>`3l ig0tGd:d! ^x.7 ɹW/xMn_ ӹ0 [&k)GE3Y}u^ [@d%I8p#Ͽ\ǵܟB!ty'FW#O:w,k"B$?U込^q|zc- =d1[^XՅ"wQ䴃"+KXRvĹlvKV͒81=&=2D\0;bi3MdVIׅg _<xwB0ERz\j(Q*d udye+BbXE%l'byli(Pn<Y{dLy3-Qv]8d(.G-9U*Jak2CN$kAfU)G\H[zP*|B TNGb(Ay2Ty<JeHUjYR Y9(V9fyeH䨞Q\@aBu1~Te<:Wj4z=QVZVդ]%XK<ߒb\.қGQ9(ɰՂN-gKi͛:5lRuafFf*W2LXy1Q _.1:;K?W0:+{kzT=,'RISYwyexSb%(ȯG5%<<ڞl{cd1G?uW5gSJ.n v[A$,2hDVy)bVRθ:n9؜?sz]$jVy,Q<ztܭ NNS 9?y[&2mwUM*tploPںN0O2M@c+܇ΰzsX >F3wkQ6i_ң?)P,wU=zcZHkcgmR U)ֳ /W1)`]l V0LIx9~OIY7BC~<m"lJQ,'sP#O]z>#j[YwⲒkxd<72xyGduH>(@,K?7 (4U6xu8WaS3Lc[#'zuP3LF]. OCyomA˯-4 H ,.l%62E0YȠPDe"+8dQroVxdia/ų]9-UUݦG޿9oL),^16ʦW:Cv=Ͱ"|UƐ nT_=!bS!{/'gFx E Q2g}z SY>M;iDǥjsp2]Wop::;$4ϓi{p (Iℜ3Q74"nIģ$)&aKb B,2i&8C_Nޜ&e&o0ar$xlgiI־>GBb{?}`)n_A+ x{r.H#F{<WGNIza^Ae6y-)8\k{a`z"DKW@,r^ۓ2K[+ڇP(`Ϸ/<8ˣe2ku oHM&Ix'iv-Br7YFWl~e;vm.2hLۃ`}@uKp>?Pw %މ9G2Qp|{GAwh( XmW().mjt$b̈A2HDQOU8< r؄Teo1>y;48L0/t5}tcZbZE_ Rc65iDk/ZO,qDք'8؏s37AY%ZJݗeHp~r%ru- "7s l]ϷKSVo9$,J3@:ꍵҁp1#sU2p63xfE{5ɦtWsG+1¸2>crW< nI1ќfpOxz.׵9<Ʈ>GG*Lܜr{e>&p-&2K87@ϳ#L#e b6˃Yt–M$GjuД9F>I&<E*MZ9W"yw<Ќ0M!1te-Mu1T&s*|SfRu)*y'fovPy3[feGRIma0kۂ0]WGyl8VHY^T i{:s$1=O" i±L>`CLՕ, -u7u5%kJUQ@ s29\0YŮ1˛J]*Ug-d N\1]"}&I6Xb $WXAE5<ݶhĝ %gh]-AtƘ YK^˶cZK:lpu|7FE'["[.eoBd:3<<óiu 4Cr]l|bQՅ]Иy8^Šp]ܓE2PD*.q"F}SZ;0YdxW}MLok`YX&ԖMɈf\;zoJm@sU%_0-,AMԺ"*gKC@aRIMv-j1\PmMeh*x{e9Ӧ6 b]`)6E:'oE-l3ϵYoCʃZmֽk+Qmgڱ",.߾je;>e2m¥2Df!ՙ2KbUFSD+f0V$ 6t$ NhlnKHIzRm6:efǶV/A4)yyEf,lOWO6>BOчyRь$ c!Ʀ!XZIn;-/ŹXG%A>,DXV2L'i.+!+obҐoD$op. x^Wzr[RaL! Waҹ 5=96ݹV-<:dYzS8dm.ߐ`TԤ!:]mHwm\{w0tH\2 V-ۺ͟ 5Mu=]8lі Dn RՃwhtm-==\ Fy_=uNYdy<=*>醱+KUT+рizEٙ*X:Cg\GL˲I!wp޲][opAz(8eu0nNLҙn u\N/'LrӜRLUy[=l4yջi 6? |t5ӪM,\ %[ĨhgmA$2D@EŸT7CHU{e΋<ͧ7(aagҒ4q~&\i-y˵\sL$PLWbXw͍lG&8==0tz{Cy_kmP r]8,{ a~.]`fo~;5AZ~Y4ʷvmzJg4 Pq{7#2;|F( G3&?d `ќvPݙ pv[܄iz0)z_e68@n(_ar<]iyh`O0|&aYM,o:(ׁs2rߕ}bP(4Ap :+;2/n (ؑX0{ B=ܥh{c8ش0ٻ޶-W^S=z/\QqkmP$e(-)RPei${ΡlOh(-@n޶I0~ PlvH %bvcBn I~0vk|(^eӸH_ŋm8=˫AX}41+7`B!gZRo)4V`X;!EFnޫnicŞ`*d1+hP :z@Vި }+3h'te%ChA4o~O4[K,~7 oi}hCVfՆƏ2>E섎[7'JͩMN>4.G(,jcrwI "v3k%h\[?L͍J}e$ Yz]ć"sJ+j6`f%T'WC zz$`'ʯk/HӣLȋ3dG ݍA*%űr֪gO|߼j)H29{ѪZ[` ]?:OߞQyRyQqČYM5ؗ5%qx:21`=3ΰ=eU{A{}taDxEWpKpBD,h:ej,A fM(]4x#A%P$<'>A upi<Axsw@kC_XRKAQh>PzD!;c&WYWZ1`4;$+c7iL: ѓ.cryq`~8O 8Tupx!^E~7L %|Fx9F-֌k9V hP;.Q.F# P75!kDBg bPgw< v!/^_K"Z|UP?-ʙRbb`ȝ0#tՙj2uJ/'8Gh.ʒPf\qXLU wIÊY fΫւhɪk}i&=dr6e2.|4C_滙7%`MܠmhP蝥TB,\vXv]M^)1Qv% Y5g*x+M==(cfouK6N]e,/ ?,awV¥YNFi<&5Dy;qv=wZZ8#αz:JOx-d֛ILz>Ѧ4"ǘEys3Sv٬4vfH>El WFj_w;ncPOUBmZgT_ -^/~k ZwZM˜:IGGIKq@Z0DʢkT8N5=>RL\o-gKM5hiǦ맶9rb 8LG^ rxnpƢM3д,@UZ,!L^XFExvtN:xI;tDmx-&F Z~|Li>a^xYYlDaJ b#Dq!9ю:#'p];M._`AZO q7 燘!NV[2u7g 2pM4{]bY޹y8K␾Z1#q?bWN+'yl?:"]NΊtV.0}`H_}c[Bߏί+,n]:K\ ВfYyMʦq|!3Fײi|b><(ƳF##_+|*̷}hvEns6h/β#/EЧlnR.ybؾnTCg~`뎹7 #l8~$YN<@3 ]c3ttG9%iqS;.l4FyR.&ݵ[z F3"a6ܪGB9juC_xݢDꭂ>BaMP[ m!) ?MP[ l!h )hẠ >M𑂆[*[*TO|B yc4&e^e]^\|G'@'[6gZurA;:X^ُU޹pcӪnUE>FN\q*r j5HG?I>O,ϣ2CGlzI]-'Ӳ&waHFVͳDT%$;DjlNǸ'--M;,b+s%~ϒŨPe1E5)9߈ 4vڙt_TշJ0&y 4ϠR<-P LL[{huÇߍBZ3,'Ԗ5SKpY=*[>xl_LuQq&합*e%6::E7vxV&CMţ ,|[/ֈvq _,۲(/gӬ'&+pR'5ړV6Z2لJƐ<,jfS×wYU}< \9\HuKG3-!7nBc ߖqz> ֎qY2JmT)3m ֣1{Z-e x,vx" v VG" zS|L @|Ad@3`3pLW[?_!{.Lj-Hld:f]G(Ow4Z_N&//'&\s"CxGUu[zga`?9-E]xXŝ5,g(2J0.,1Ғ2ML zL.ܵ^Ypzl/wWJS\Dz/{2;ˉ+N*A,N%eLQ$aZH_&( sPD>O SY&Z勖Keޒ%4@[[ CboƼ.T珦9' rAt-ŐOe,+ c"QhbVxA䈳])%T֋+.c >;2Fhf Nxܱ:~–4)aDkf4{-V,_"'8NғLAڥ Q}GߜW6 e}4OlNyh_ &jo_oU-{z]Q; Tݞu{:C?̭zb'> Ui_ߙvhgpujtwh3M6v߾gO_9N+PrkzxMg8CԾK|:M[`Ob@dWc_o,US"+ZCmԟƺ+fWSFtF{k|=HN1l>Cz}6Y0[ R+P6b\;l*KJ ɫnwpUq)'ZX!=zZZ 9+K ^ݩ 5Vprjh𣭋 1hM5vwlDxװ?TV]Nq~KQb,7o4Bڷ&%aVȊT:J>풵Uyt] bAZ['@OfeRWݶp[ޓ.miR]ȿFw\&J=mg.5$O^wEKɮγǁUHKbEGamHm@W6Dd(ܖ}K{[+a.Θ:GU_WVzk19xʆyPsppr-c-Ǜ S?UG>+r _MK0EV +DU ť,-J@*OeNH׿Mk Ey8,Cue ?Tq; ͠2]2xIx~ aZ(&Iǧ*˦ꕛ/#__"WrU6,7UF8geV$OyF4,Ekl;I-L!N]7܍+\l!9˖K$+E}ٟM '|C1Znv"=|=v+;dȅߑrOvsvN2ϪȻ8W|vuaV 1@įwVp䎲V @"tȪؿ}ғhg*~ym3L8ݡs̻KΜ-fίjz?-ܝݜJ?]adO4Uzsӷc:0wL~X˹}]d':CD+ۓ_MI>0t1Όʘh\[cO_wֿה6'*YЧuIjr;["~W7/YmkuWߴ_7u/qݛzv eҳ{{bC_E}=Nߺ{ܗ_px{s54 /̀6'XcK{ghs??Twq/Esx/3;Or6We#%ZXcn^uEN?6)1SMÎL±C"G(cCQǶ>t>1v΃߮&GoJ\~)\P޾:x{Dͦ ?G3K"Q8p{xqmfZ' T!"2F g{Z`j)P<'u`B.Kݷ?, vN[t//n:9#KpQUC`CwUAa? \ aRGr} $eUp=e 7WբPa$# v~lp4ÂpGxm/aJX9?r/Љ\M1c]\10ZCmm=c+1 i:*5(pFҙ×L/$8Wެs} XxOD+81p|z0LZe Vw2(cĐ{0^Ey0'D91uߞIX^15!m⎕,DK+Fl#FG:7л8sUfQfbG"i4cİ,>Y_;U+BH FCu%}BHz}I B{Ҧ?ELü/Ţ2G#JEzIg٭Gg[xۨueF o3  X6 ңFHq$nҘu@.%1 1t.F$VkZ9 B0,36愶I+CS f,l4Е¢W@|a[#(b+hpa 4ѳ*'Oy#SQ%Í(iHbԇc+e){ϙ(V#<6^.*z:è3~ T, 4K.SM„k+uLkF(>( _d$jMG/$HBbj3]c a(* % R+UT"<0Ύǭ}"ь=>%=ҥd@ R2FepDa/جw! ,hЁ Bue DH:S|(pL\6 g\f)k\j7Űnɝ۬4Tލ3je1K=,.ĥۏ3}Sf8!Iѳfȫ`<㊲_tn}Cht@ef9={/xT-2(Whj\*-B 0V|-&oߚ-a)K=}gr ˹qB,q,Q$ީ7cytJpKپA=og^Lnf 3sx@ E`I 2 $t&nphml a&/eiEĎ_Xq[HĠ̢Fa|}nG56R̫Qt|Q~~6tzp=+-`tʽ8r!&70 5%m1I3OԳMci Oÿd==%FZj2|R $pCjk&s :6JCۢheID\ oF`$O . {Wo}f>,3gz_B>dqG+rftҁJდ_4<:`G@\ F}h S`F|M  AnXZsAH0vPzCd {-簌cnS*FdkNʹB&V4FC$8HtHءVÊ$Sc昕1#L\4UéyW[ߎ \a_Cbj#ܛ^F,,53!YVA\d$o}ޮzH@nQ{ΫJEwR@<HI/7fZL+jc:;ʞ쒷*F YaE /Wڐ[. w?DcuZ8W8GX3 {gJ:x&J9ud* L!=a&7jp'HDjOPeْU=N\sax+Y%r5c!ߖDhorI =o6\rn4H˪͆ſKf~ KoNGsBr:`@<ݲKm' Y7V6jQ'&uƜ~cGSP`DY`"r=k%e [049V=8eM05(R9'RB:{&Ѿ΅\3_OvQgS0\Q7ԑ)3?Y~qM;z~TƼ(&W֘Q1G/ \ђ94V<[w 3+p ]?x+VLՍ>kct彬q!u:j@\J>k]P*=fTy,&X .'0FLKTz[.@¹/ܛ/6miczfzqccY=z`Ndsx2Yk퉒'YuMvFv'X;9֙y{F } ݃8ɤ6!REA`$U@PPв٦B̤९|g5 [*C_`c7]{Pecr oטX"ƮS 7c$ۗG2!hAy\TaiL),>]_QbiĆy@ &S'߹Ȭ6$U2-pVfS;F4ɧ'Ob< JsxI!QgkԵˍG+22aQtUqmWn+[֓Aހ2arf9BLy~{"qޢKW%}0л7x=|f66Co.p&e='Uyc'0|H1kE'E5F1=oŒ8+s0ϪeQΜ4[ќK F݀O^#kEwЙj8uL"+KN [I{GH"ın |S%nScP}c. l Bjސ8ڞ*djTM(c-]!Zƈ?G!hmZ 33ʘ0t>Rln0_ʥ۷g=0xs3'u0k$޶ScA&zLfhy#SJ;&H3NeFoUm6FQĹvi Ç=7jΌ-Z## y;C2ִ8W?<x˛`;U/Q7+?r5"mS_#/Fir;8^mMwL*5k7-Ksi] %LYB4 KXLKs*2^Fm߿s32!rx5%YYTc7s#8n{%G)k8]#Hk%%O1\9FPLvjc@Yfc }ٌ..e< r5tJvթ,.ϝ":,sBup|!UAQ{תk={ -~{&zȅ8ΔW]KDYrg2K -~YxO@ _O/vP /R؜D&so4{3k1vdr ݺ((RM14a Q?D (g7BSy8R32v$)JyJD(pxBoz=*./cհpPY{z@fm=z |SήU7?FX}M|~ugta <=MFΟǁ3<~ Z諆(͚2x=?p>ӆLB2eTH o^}Gz$xk`O1]@  }0jqh~Rw xGA0ɫ^ tEy?kz7ey2;/Ãhp`"c4~Qġ7}p؎/#ȗ_ fO6U}E1U㇑`'e+-fqQ~zC.'6 %jBq5-+<ތFQV/flܥDdӲ}B{`m:Mj#F:YT[2~vG/Qo0R׹SHM~TDM !~ Ō'4_ Jػu,D1=V0:ìbE(z#Ii2~`ڱ"AÜW *#DZ镝5H d4JrLhxQiO*=U^UG@K#Y(Kyuk?9gm6_#0Su;eΌMFs^x>ua_ 6 i{drC%TfQȹ?~jou`1NQl֔S77@7o$qpW 4DL,<94gѰ1jֻ. 0ŬJįieٯl_Ǎv.wٌE~AFy0q:9'!.G h bi;Ф~L/f6b ӂ~8J6m?WV3U!B$]"OS(\9/.LFE~ҵ#<֭v,YJrFB`Ff?~3,^B7Elnԯs˕hrs}W˱>a$knzh{>Mృ 𜠝( :kޙlx;OmiV&1D#0#+!̇Tlnr_k읯o7czk3EUPNQtD;Mgmt\:ȷ6Eq;nuufwTN)Y-% vH(I(E؆⨶\wfM(nЄMnnIK@yq͒;6WLÿ JtK[(ͺ-t!$ @0PnudNIaC@%چ'a0,w4@Kǻ^Rc,gmY~qQ8B\#@ r7JTnA#-+:54bO6{;)QsOyNCc.%  .hgnM IקV^Bz㒊.Cy1iM8uDh*;@;pj3!')͓ +3V`3Te"mN05<_r^"sϿZ8v9Kv!i%TLFK&Dc(E4(xg;[ڂ fRi &N^k#wX0+pJ M3pO (GO жuL%JgU`QUk-50*-[hNk(x&W' ad@KRh4JےENmkZhgqQmy[AE %IY XtQrYGe4FHJc~ bq)˕㸓UQݎ+} n<6KiJfVBoLT\Ao &4z/0#pÕ^⻺si%HgX "Y~ܒ AS[Ab9s50$XSд%:!͍bw|r-&NY/I`!a!<[I`0L`е=:!',. *aȨ h!IA0s`` '6bSA P)7_֡&ʜu7\ºl[\ϵ@ K=ϫs >H-I N JIn2Qbp.0'Î.̷Fn4SMS!;6j]ncg 5:"gr**'f]9Ls Օ:R6s[{sBzYe1W&8GVQc4BcAMu~U-(l@Onç]DrlqAdlNrSsArnBnD*Ici& 5X9\#ehP6XsjlKhYXXcd%4<#p ѣ >vaf;`Zʼn2Fcƃj?O/a$/Y_.BZfd{6R;gVz%d#+͹Gm('g & جSQ5J- ;%o % 5;4~SKV E%}{A,LNK}4k3p)9v[B+HǗmGqk&KV#ׇ%Rs,d'R~/C~(^D,nb ַ>'xвΩ8L4e(GZ)=`{\fZhR #OU.! [wvmqd'NN۩^?>f6Ɍ;'(2a=IsA}}]׋xOΓwEVC8_ ʑM3+5gyop8`!9QǺ1rp%r/.dW~T{Fpn@݀PξxG/2x-C[håT@`UV2|/7R`$ivhk10)0잕2u1X{R"ۖ6g8̒hF0RNFl;TY-cLa&`be[n(ȻxaUj}뛡ߥ@[M[-ߵ"L*MWΔSJ[O8*e]-"b2P+6R;qz;< #cZ- W,(e;ߜ`z;W}ºc%FF<'콰v_qL)JWm>r59S XcOL.(]2Nbg˳Zf%&/ eLo^dvIBt1հvv^ L`:/ E9;!2b9iMhr#ō ;dxwLA[[JՌ\_NzOxqmRk$o9Yo)wAnﲼ=^RZ6X|@qg05|9;b8r[4bg_[Yo_+@B9DfXB< ~ $E?y?o}0VUsx k6h0 uCf:u[};!U@P` ZI#e4;8QiWLKva%a(uY!?PghP-nb Xw^hE}e`+K^{KP {Ձ*Y.͈Ox0W_Xۅ*|LH8}8Oy#hDtK&\AХoHtvf$7X ^UwM߼ f¦y'e I K_ ѽAQPFgHP8 n\([vca&%YX/ B$u~}"Yq%4QN5]@ߕaGs,i[D#Ol>#'(>;DI^fJ)y iK>?n{>?P Z$Fw-FƉRp3ФnZKڢ{N( ЏM[%# 9]:(o<|2(!{q76ViuDޙXhp@#,DO9r\ؓ`]}~Vi'CR׊o-~o[ߩgKpsUHpKQb{ûVtkMJb~Eo fZ@*F˗0et,ń/>Ԧ̍^ qnqryhZG~9_yU> oӋ(42"M`/ܕX (eyׅlR7^E"f6T݅:QPg![4#^6<cTop3A6!@71ˇY}JD&$|jE.`X7zy4rqb0g =*rfbr~k%Y Y6 '۶{U7q{<*s$ âl><; ?ϔ./.#Iz^}n4 DOҪv[_y y2gmJNXK <"&W]ҩW6tSQG2jbeM8;}q^KOW|\ voh#cB! co0VB5 }}?˞=['?5oc೑>&w7~>87=/G«ymI/#< zW.T`aOޏQ^|x3~_|J> :T/Sr=o /b~a>.5$yv\%t9 &O:sGMYיdp]Gito!c_p%7iEJ"bF4ׂoIlS8F9+cgQl9>0^SPp5J9(?lE.MA"x<kl:tE!џE(@; |7]7ƭ*t9Rߜrn~]ϣcĠRażUFVrBTRXWYiqL*l`6?ef8l&;Y̑YB2ˋn%ʓtq2[5x9ب5Fx:͞*CW a4z,(Ifutz}NCO<3pχLLyFhXق,XXפ2hg>v`Zsw)^ˬ~tڷZzk@* VL'ST{2՞L'ST{2՞L14T郙jp|*!4D#ƜEi[,RIhu^-}`r"*qNhG ,NkΰgJF,J h1»44JؓR1X3 M6]o,lg)6& ]}2?"L)^&[/& x4_lɢ_rOzWrc;`NH$Va,$)UBba[eu98`\J8QmI1>fSAU*R%LIhL1&IYkǴ҄OVZ[+m\ۈ΃HWCa ]>r}K):~a̚Z}ajmms_][F+&ة " C&0j+QKIm^SŦZbwT  bUS߹~WY nj٧3<+}`Fɕ0:z&|B;d`skHd"tJDĂ)1Z&m=A$3PAw2 ;Ќ1&("T`;Aoǝr|~0ʣf=趾Cg˼GpVޛ k:u?4Z?4|Gvb߄s/u4RL&̔qQ`uLЎ&ME=9DKH@0; ]\8&iTTcB`/fn!PIWdIZhⵞ4_n6A|>h@~ozou mwc6dD&J!r  dMO ؜ ]'"qTQ[ uƞՃPLJM˽z/ԮԽΡCպ&(7oA[7292UKƌL.T*ͪYzlRaݺ=rAXo+\pj+|HS{)ch*n9|{ Db% B2ğxG^_d՝oyم#7;4VD9 8xlDZ30 }%\Rݼtך3_v?7vzB~zŬelzg Vׯ+ݪ/ |qYnׯ_\=N?i٨s_( ?=bgM "5wFή畾J (X,4v.v >fyhs'VF}ls3w]kdMdMdMdMܿ9ǔd*{0T:Ng*{^5Dr1NeST:NeST:NST:T:Neg+uz6N:R!/q ky9'Y\W*)_/2Z 1v'\5bvC^zV[-cp^ٙqYy_ڊ<3?fUEvvLݭ捷o2U w{<9٣x|vfovG뫫+̛ג{}UWW-;,-G|)!9υFU霄8w]adw]>v? ;f4;.6/;7+1*wڔMzTu\]^߼l& KXK_2`UkaE}f?ԋ8{oZy|W8f4% ֠LrS1); fw(d?@ahaLk_<œLS .s&G[{ ;HѼQMx,q_F3w}O1">r-Vgᝠeaf:9 moq&8G|<ķ\+|+ݩk1gJOeZSlpuU~G5;^yfaޮvp4폳-%yw :κݡ˶gvz+=_p?|u R;-QS;HT )*tgO7:fo;bvyÏgUOU*S{^ej2WګLU$wZGÛ^-wTr(E+dߗ'Y1l}9ͳJ.q$.BBX9%5D5%DiP=5lmnMc8$D-hD`?kM" ]6/A.k_ξs*/8y]~_TJ +1e@#ErwHd2[{ڽ<ْ=T%XH\W q,2#`6Y{<ڢnm5CZV|3 %(+S$Ǥr#$P hhx_5,Cm9a('Eۈ8S 95JͩAԚwOڒZ-ޓ~o0--eR%Ȁf %bmuZW|P S:`@pW[ABLc+} Kr5CT@(~QUˋO9] x1:8@"UBt̫׸3y)ޓQEݜc6$%BN7,R$ ͂r> da0p\{"y$gY#=]7& M)o+ $1p )t\K2 f}ijܴ/t mn#ơgq)\/"'@.ug/>>ہX27torm,ڮ~r]lˁU!A‹+ 73Q1eݟ|O䒇[D.ޓjv;_̷J~̱RpcH2A1Z9bnhx_xd^{_I"`_?hF! X[tsbбF h}I`$_U84wcv.)A=);!T@@48ĸ1`fu-R KjFwZFFsM^QIy<"rDVzXSb_XX|M,hpW,lC-;4E] b9x>ȘAbw\ Q#ΕޢzcnūNo蘨'gb9,X9 +sHZ#cEJL :^JhPg<nt-B%5"T4TvMDH Xxv}7AS.Pt ,|ܢHJ&c?..x,SLleLz1B0vQ#G*y}g8vGL WuQYFh(h(ޗ__Y3(v8pHA"ΈΪyWNK:hwm]4/◟>l(7fEOGP|plVk4Us\%cpK&Q[;y֘L5zߵ)BXȪ5鈛W)EŘZh6d4xXs 9wCˁ Og4.yt1>9=o G11AxramUҘԱF..ޗ}B."[E'q|N@£(E Eea JFsB>!%.R}m2S/+wEEs IN8tӢ]όU}a,\) u1Ư7[3TF+Yȁ!2D 1eRbtj`vMuĖdDŕ(nf@ңqQǘ{ox `2HU`̭xD32+ǡǒڊhx?jE.wmR dEZX ?P`hx_4,'yױƈO@;b/*B=1P`?tpD@#\3Rƍ+*=\* j.T[N $5,rfl_HRsdzkjFjA4hv7bFߺrOxE]iFep:n%$2Ma`ē`k+=F0,A&3uB&g sGcA5o]hR6xI5UbJrĺ[~t(EnayV%(Tv ݌s_-ese{ia˗tr%vfI⼋sJocYܰA7, V(5j :TuQHW_;z% 9٣Z.3l DwQ1 U#njJٱeժ-\˯Í1?6װM 7S;zơjP>J>ؿA1VEyUW{*zϴG $ 7qgEF0^n* ~"K/eǻ!Ybi$dESMRd&Q4r=81FCߺjT9X=hꀵ(!ض舥.0Afl\vbSt:v.*p0QQ.`M63p%j\5*AM)c]G7v1o]߈3<blbw~Q#])p  9{Q0 f'3Ds.#f:ǜ})|8Dfwg3fN :[Bu#IkcHM0mR-Y_l nl T9{7>3:R$"AcP_bnhbcctccP.U =Σj̚.8˭vv/d4B(vqYDDllQ B3"fQQ%'gK%QQa:$ёd7%Vшw Y%&*T*"l:ӆ1DW@ߺ_9^}2D`DqvXDͪ3.a,=TxC];#Ͱޚ#>2(B9Jd4r5rsSXqt:w!^ 4Φ4xf\FIF:rIP#zeUeiC<3.iئ` (݁Sx x;h)~y݃cIߊg.6ـ>P"1\All#TPq*i 1RL"m}n([yD/9*b3@5<1V J3e'[~b%].5-^.˼({Ɵ5`kƣƟť͒|cмLP]w=tP)7ƞ,yA1z~(ڬ7*w@QI_5x 3bC3p!Np"ś4DL73A&JJ;iq<~}.ʍ<nG6# \ٌGLԀ0ŵ@U"HCEAw8Z1 |̼74hBEyo,5jo M7v"9'#Y Z&\}ZA#[W|jeY/42>_+\D(¾ Zw_#"l= DX"쭝k{yxGL (>b%Zuy<4aܡ$ Kh!P!8>C GCA08NKƜ{:(u3Dz)wl8 N K+.wTkޅ4&9>9gXS,O5K;b5EQ40*֩0:Eyp@&5+%$ a#IFPVYГ|7_.ϻp;*bp4TD_w'0 h`EAV\t:+IO CjG/ PJzKDxաR}Yڬ vŒې |QԬPO N}.,b #yW(YMP;2boȈ#!z4 9rQeQN3GZYCV cYn=IX}R7UnzIBFq[l-j'/?`MqF\sEt {N0 ƋJbp#>3NxMGx~S\қY|)H8PnMAtM 0OYHX`D$" B -![[ x!q_ *R(eϿP+UZ{pz)w{i`eR8a6Oheͤ OB?whއ 2`ROc EYӯnK譥֡'F*,fabc4QDDФgsP4g|pdXV˳|woJoKXEb˵+`v(,ϺXXbq4,rٟW.d*kmq7lZgpW_.#pџ5e{bq-&C݀.A2}ސ+`% eqȟݯݏ?N,|q7͠ή[||Xʭy~Ⱦ͂|]N(X@vK'沵b' z]윟~ڼ]rWy^~yVˍno?nrS}5yc%](;+ɛ_YlmMԗ%@V™mUI|ӄsJoN-eYḿuQ_`_ԣ3E 0vZd,:b&>~*6E5^CkU0 0VŮ1-G;*+._Oh#>OczcѨ>/Č`$ItI{^B6t"#};FHu,"NXf4LF 0 nE`*4rN|MЌދ>|2zS08d߹($4LG&" O*:Hʮ5\;4f,qsz:^eHUpfN?;[2v|8>`16`ܛw9(v_2Z2#€:SY@AEז3+f"/x$cİPkV\@_Y`cѐZ'qJ5zDkG 0V䩧1Clj}ܒG`SOtOqkJSAE2(MG< x e$Ly`a=.y4p;1l=]80 ŇY%\F=n>0׭g39a6[\c/2ЏעD捿5b+U4r'VѨds9DX!rBY#q#AUcV:{+e٥W*|-{~ӣ,'nW1=-[ܴ̂zlACs yp@2-F(*|Y.f6<\. :^lEwGMk[ Nhʷfdgp] VU\L͒8Ȓ$q•/̣Ŀ~8^]-dnܛ,\~XZWZ[Uݧ|ɾv7w秼4DL7ID֌֤X n*xjVпU}6NmظJץ3-R_88\~zͨ"Xy,<p~pxq]]..}5⪏,.N&ެ6+DzSΖˢ]md Wyqny]s| "b-jan[q"VsY (vϫ0Ovx6YSFG[Jx/,~b,*EU—z9HŒyyCh`_Y+ݠE?6ڪڒM0>&.NFVBR0L{n$j^+cᲹKRI.OIl[cYH{Wmu[huR&eH{ch4D>PRaLp/U/yyG6w5j@+ZB59 RU[eА5n@YknZ' 7oHTZ]I3o!~3|-s4zlg|ZD^̘b (f Bi~a>mäPɤt@:eF .u1cm-D@] (\ֈSׇ{j`D#h4h"W2K_nC ;jeyg +*ϛ*Ct+[A:xpP.طxިr1#O3?=r,燜+!X+-=Oɮ7\!O5&Oߨu2JYΟ*Ux|BG (ro*+E־o0 @uc oRJ?⛣Áev"K}y5ڥގ7@;6 ƺi\ť2WrݑƘ=`33P`߁ւeDr{+>{ǵO*O\(XNjC V ٶ c Uk)'r7{#a(1oPn(C* o_U&_g*c+ۙ D>y,+ftӸiɕiA+ik*'JKii AhTbԲ/beSPlASGgl>k5xuK!J]5k E Ƌvə5R^~Zm߂Ay׮J[UVѹwmE6c[oy!N + TE?wC Aw>"C0x7s3Ox/~dv#t WYEnmm@Q-CcF-9̓f~7"K)ϘƂj Xy}XQNW6NLr;+~k:@oqkFMR2/;(`Lb˶b0NZͶ,ޔ[A.Rv|WkWi&kk-ULM]4di)i^ e9KO=Ю.T-,o"*ʲzB]Uv aUccQ8iK]D79FCCE.rY($€qv],NyJ7^ҺދmCĸCEl}VwuY=uD۪"|U&L:Ueˮ8@#h4QÇu^\߂C{gJھ121WBfuv&D0' ]PS+wA.>^ImJq0@4j@s9:u= TBߔpQm|@ *\w̶ D⡏B"X-+#v-8UtR/6nZzA_7>yF_=_O5i%Ee]~##4P~agT?rZ} q}>:yAs.$ZXbE|amELYm!'M‡2iYM``WW[H[-{ u֣&|_ Oy/sū9\O-ݫsVUdMdI)TH/J ,9µrzwo{-c}؂kpNֱg֊x 6YKe4ւiA _pBkgKZHG:Wd}c0l>\}jvVÖ^ALl>>;Ȏ5qd=OBI'.Vu9:+ Mp7TH5Uwz\X{x Ke3i^fx@c( NqFIRZ˽ KHv=Y--@RK){oyy~x=-WQ̇s"P^0#URBӬ,.;M89:;~@mq<7#h^9r`>؜'|xPPRp筍=zRh ^ F0N~X"e Yڸݺ&1|pYLg̯ s؜++4tQ,噬l[tq5%B"vZyUM8uU%<ڭK]Ẍ́a M&a$T6UIFX<~J+jhMLֳ26ސm aX?Svh[H:S2Wtzy{Tc>p3"{f!q6x_Ԕ%d s#nGH7:hxgO0 <"Pm?X;$|@ 5z2'ǧz57LW0>}䏫շ|{zZ ds'1HۧO >|~ϳi9N{vicc 㖸 Ly1H[*wv66䰠-D׊ѯ kf0b;kZ].u *0КDlFϧIH?-@Ӱ(6M8PF"Qku ?$/?0 \Xm8V @_qO1dt/oFӤovt^q;r:Fas:F?K.Ձ{ğlY#vc~qNI[`:bČx#|nĎgs~ڽVa17[GQ)Ka9xpuցhd98bӿ=?YOV7 ߀iamluT '\PBiN,|Mqd *N?\l썲{Ʌ# PCSU5\zVӻ/f-W]YpFŜ[8Y()"㟘M G4勋[R +1p,s|j"}+Vq\ Q萝x$.֞uy[{{05?`H" l*%U8Mn^Q5~5G4hzY<7 @?$۽}Y3?wM{ybݏڷ:E?: u p;3:RtAtNk%l,MCQӹ_FKG|KJ"[.R3(wzbggT{r0txoӦM+Tkm[D|9|tu\DA )\ {sU:D $el4#p5:xmqpH&)p5=H) (?*)÷&)X,֫5#Ca9q5&M E zkp@ HO,t>~ +P!X #إcM.?kK-0}I){r`^cqRHB]wXRa;IyT쩃GbX}=K(n ~Tީ>pCVc< g`EY8QND F+-YKcQ 5N pH8B{-%T(WqPvE֓2BeUoV& 2w)#;?Y٘T)EW 0KXj#pA 0(id WnJX[e%h+4K8M-uX8rxjhρAX8B^sȥ]' dkb-/$ցN<WV0TKU:,#{EhN,@(f(KPnZR.Ιɮ PQ<ʀp9\p PCSځ*_%)N7gѺt#9D%0Yڍb l綠|-k;r++y#Nx01Wƌ2X tE!i̎Gb2{Ɠ㣑 !R_:x$N"º ^ZUbixۑa%l2 GNY$JXe7#F C)= y0V:-'cOE~nR!J+ܗP/}-rj]pI_wT] BIKr_<֫^$(W1Y$W3AK X8#Q?>XWvbe2 GnF CE:x$N ֯=*1#/[~2##1H,eYKY#(PՉ"])4J8.lETy) 6n#W]){ڢbATJr7lIdKrvMy̡8p>\z<4 c!<v$/ePc}S]/MQ2CQx;xQX+yULӃ݁WQ]' ;* m4ߣla'!Vg7XL5ωDB2_ƥ9 [=XW-%ݵg%mzVXz4FP޳F0L^i#pcG9GU_S<=\.t/&>FL¦Rvq؁cCD2|'S>6R"UYg.L+*^V.wE~/9ث4L(!}imj (Pٸ8[IXӄl j tT(n1Y\O^2L\܂y)l. Rvۭ3_/PǼ@e~!?JFɗ1dA(c|L<8^gy1n CByyQg-AzK[q'wUMo}N{LpʘSG)mF]?2ˈWȒ9ϨԾё#ZuTGD(q@r$edR 5ku\nE'+;.HH|9+e8cu~KPy MDi8g< HV|sA'^Ig-;sDK$ϽfytMyrW׉. n3NQ0|} f"@Qj#CM(Jh(_MQ04AQNQ_vч 4E+E1jq ʰuA(pE׶uQBe,]c:Q'5#ƙV4TޞH즉2fHJ@ݿ*:`o0Xgno`Y~=I IPDEkɮlq}cja{ c4ǔTbh6"޾wRЊxk⡀c4޻gCɈFuFt'Uc߸Wj)߅Mv6l2UU U~`KL0s>ڹ!Wɇ~|)J6䊊<5Q]TܞT|EvI&)y͗S3C]«7uc3Wy^~Qj^|bΓ;{{ggFl1M/(Nw@JWͬ#HȀZ?*_r2\p1I(Iٰ*`zwE> ]94f[Bvz26~"ens&~;)ǀL=H4* ?oDy"޿ema /0'mKoQMh0z Y>KPL_&|0ײI 84G{nc&: 1j⣀4 8ϢVrlnFG{1(i GrZԲ1&܊5 QzF^lzvoRU d,QUەc8 ~ ?(c±O6'z j3¦3~rcQDQ#rcc}Wܕa"(PP@m?wRpr Fi.nO+!K;@/xG1y~<9u򑯬\EG2"t pD',}2K'9* !pDLJx6++1kVV8&% ǣ(|[ܡ)ZAp+\ 04BQ#^{!7j*>EEp7'p }E^tg: ׆Y S؍0w|Qj$jb<; 3zM/K%Sq>ށV1si T=cx* -2EG;{i|g^u \B+@n<7g-wM= _@9y &cݼg#xJl>P ,2G1G{=-Z.arxl,LV|c"IчRBk20C%\'Q{Iv̈+v˓\hRg04-)Lc_jJҭmqO׺ _ bۮE(=‹J{ >⃄&Y᳅`Gk .^w`TO8r=3cz -T^PP3NP8bB> 0SQd=$`L\ K0WL(`9 E~n& +A ,+æ6fN^!Vk1lJ iwyp[F\&dDI HR;$bmi Lܥ]}#~88d*+ߧ:᎐mlhNMp!Fo\s隣lu"!#&G>+3rل~I4dQHnb*23]6P狥wxNJM#O[-Uɺg*`2UrఅRڰG { p:ˮz¤t)>-lj@B²f0N>[RaQ0kW.p9]yWxY^~fr XsoJ`JIj ~R\y!VѪEn |c~}?w]&xb⠚;O^6Lnۯa?9q 8[b6+9Zragw 5:w8`:;Ɲ'0̚˵W?өd[GoAb#y0uQ,ˋ sJȀ{zv lwy#0x>8~HX i'RL}x_WE`l*b;3w0 x^`+M3OnlW$E/+8wZ2_|ηY 9Wnb^7k8lҁ?nnQZQkV16{Kx2c3ui|&؈Ngfk2_De䈩>>[G|5QjG%5 έ|6|KxJ1) SaBbf91D7*/,3bSʇ:6jCL)#x~?xT0laRa_e 2M&r2x/*3au*mr@*e4JY1ƓOhL2I8_ "MF/>:_ї&y/yDžhxQ nc}ܗ]IO705$.|bbר/}"I `sr{vVv 5n pg]8#cZ[vO7z&pHDcU]ܧ)=*}';g1v{LLHڒ1a1b'*+BD:렀x1/B(n* *Y+HRpKi]'y\AHWuW8E?Ӝ& "Ө>aB' )A 1&E<wx3*C{De#߶ՖW۠''ёiBR-ql~Jb}RU 90ؓ1N*վyeʸNƔF/9cXJ7+-G  Q hL[`4?SYD1aCË~ _XIUuS8zcu*k_5@O%&ap(;D#2iɭnl6Z|Ec_ZNFtb;/ J`8+9Kl墭*rc抽rEMtC|O-XiKcHcEҀq‚XXy:s: m#&6 yb@{Lhӊ0V {:M)(`^*%ES΁;u"oM|gN=\N<Рxt(o7ekg#NS8nML#&Sf- S0Y JLJ"R'wM]fdMjr-$KW'QD'5ys6mehO 8緛^hovyեw%%Ш{?M\Lz΢E*eVAUp9 sJkT]<-[j8~YPSTmu㤳w|;Ύ幼U%cޭ[mxG/^ZCG^ a|=#p }v˾닚[K.,4Yli[-2/Mh22TXƧˊ~.+Ͻ,+-|$˸ۙTg\pF5.ck(?LT5!}viqN=-N=-N=-N=-N=-N=-N[$*N=-F5Cw~3[EH~W}XK)Ŵ88M☤  }C$R[0O&abVdgf1Z=*अ>nWзJf/.ku>=JO\cCQ!Zi:Fר:lxSAWo@^^LR2U~O?I LRVk{C}! &s m}Ӑo 0Bڇ1w5$TdTaSýe..W%g3t(KS]}b Ӭ?%RF6HqbF|}X1}3`n 0D>gOT:G.6dSʟ50U̎S]{4*׾MB$*aHbx7y?Aȍ*t\[d7Wz;;]<Eq?ȵA{>ܞTSwqw #Lǂhr",#zkd*Rr0qp"nҕh.;)*a<=iFAU+ ZG' O%`QAN&N1FAJ'd=&ɲ"iQ3b72oWvŊ[)4\գhcylP]{oF*Bp~? w [vce+ɞ,ݷdI%rDJGY1&fu=~Uݬ:"+J4 ^Xq6ѭ,glBՏY2٬ ?ZVf-obYMoԦIr|LvZAzטF+Gc^ rd>};] r {߾usY VeaE ʝ$czPgdqDc7)A֒bG^/6fԼFk#lFޗgĸ:٬J[9`7oR?0x+0-t~!GպzyNq.dotSQzwWfbGTR׽h5ҧoU1⩪m5V&*er5Րܶue-=q|{ձJ#׌q >aN!2s_o q68WS[\/MŅ9]bliGMu㍮;aHZK'츳xdcKTqs;OZkؤʈuk~6 އ l6(:$ncF鬚1hjK>ziuTJڪIILĔa!E-szMEnwT@[{b<.2*@َxe-4Å$·=Je!jKV% A+t{f^S>$)kIbAzx(O_(!sEӞFsnː=N*sys)̍p4NX= g1zf^ihូxeB:s^+ǍՌ3s (D͵K+,K 1VA$ Fa0BcCeݫ'zEsIAakA\4Ar2!"d2HJ5F>$~O77G`2 <ɀpNȩ¡1zf8\gT Ҥ* O+Nbr@cpL) pUq3Y4q i"{f&' tM&z #܍B.'I"L/UcEY]Σ3E>itXD1| ]uprߏ[m5;NLo&-QNZZ(\`<* M rӱ)T.Kj9`~Z6V4{~yy@␧ <{V:1yA_Ni?StYfE  H_l0hlk.V\`~80 sنhyJgt[>/MKyAoS%5?t';mVSi`m'5:)_7H$x#jᐼEFR*F 0"q&%\eT !RT3FQ'< aw6kqL%Fa!\Cega;{_P!FDl]CNĶa&+ǂgE/V:?>p}Ճn}N^zͿ޿/" mMqq_FeP??Kϔc3`: qeX>燸I9\EY9?q h{ un>1/YG8oFv6i:q0~]z7YG ئBV?_l'; 憐5۾^=aq<|fL <3H_&\MA^~-uKA,~mh?glj?]OzW 5T`]$vㄹsp M,J@lr0kQ|b2?F^|˷͹Tó!ϏeB2~7OQNEU -\M~~ |R⾅X.u HzViWa;FiNǕGz~ғyA3Wtciu$wNesޝ3;=i\vlKzROvz=ZI1]-ֈ{~+O]-z =侁/ʰ7"[Ы!пG l^0Rheݢs/*[nT+=Yʴz3Y$ R^(ſV>ǟCW+BCO9Sцhh?⡶ yĉU|KmfL&!8a(jp36oG#pуN&w%q{8|0[o$ T$n?savs}M?r\N<˄? "/-|i,06x3E?"͍Z\Kܼ/P||iQ44NWA`EX"ۃWml *כŃԒg1_u$<vsd0KMBPl*t6Pvg9+CHd|ILBX|DZQ&֜H'S 9C֍v./>c|tMjBHa-,h ZƠSV dm-DފshWCZKOiuN3u2|hq4/6Ls6P̜f;Ƀ&hk(7|L+݅ /veߐ%!N ϫ3%gQ2j/ׂB{@QYۅH\tcoHnj&KF1I*‚4eE'42JX\1ԕCu&iyѴ]4OM;k A):Q*0K5,=N!rYZ78Ca4g2=_?԰n`)( Ɠh?x0hgp5uv8Wrҗkxu3][C}P:;_;@n.Ԝߓ BMw_.0&2ᰪu{Vt*㹖0C:f"EQ$?򑔚ih~9ΥHd4KQ* HLPψeYu -BkZ}sd|E-ϽU %ENTgOؘ-= ۆQJw.ЍU 0"e Ac,XHaL.;<&VM*TMe{#7RN Rr=8Be2j$FY$:`鱶6aϬ~q>]?LO/_d_} 6_mJcһO䴳FFC4:᭶&H\Ҟ-Ĵ$;d9sAuRMl|lk/'x!%f3XBv3- + D"\'gFY$&DHS9ᎈx~k@EV6;ǡ.<,-?8Fjr1 !6, Gu$j!>qԉ>GɆ$#V&Do|t@Lh*q rQ;D!RaxeIP׉0<[кbu|b~d}l^ !"8u_>Ď2VoF5˲F>]\Ut΄3T s1*a,aVr/p+!7NR3)G+| 4O.S%#U0Tq҄[Qт"DD%ݫbrm(aY}y>FKYMH%amMgE fFxm*c{!^$?<i?kjH10SE1C;4!:jp+zJ s 6&5`@v.ttEZ_zKNn;|IZ;(MvPҢOF719)OHdOrɯl9=4U;qw ?dz?W 8|>lw"Lfcp6NAѲQofVK,a<"3kȧ4rf~\775 nn'y;%'!}nn 6hˋvT+^ |oj0~>{r.;R32cVv-f<ʕ(Ξ愃z_޳6nk+B{O4-fڻ|)#Zv2i1$+VbJ[^[EƗUjTqsqFhSp+#6dv|S0JӅ]n8tt4_!N[`/x{ \>+|L.`91%ͽ OvgDrD7fX9;L7 ig2):N}ZB"b5'39ɰHH^Xd ݭdg$s6ԉCۼLgH@Y$X*lNK_o3ȯli^mZFW[[)a6i Awće+!@kkKpמ_55:S͟ []3l/$>ԈZy 5(8@<-}*K`eKڌ~*8kA>0ݧ@aSZxD3&fۯ>F4oU(G6j:< ,!}%Ag0eدe5gEzԴoF3D oakJ@Qp ڙA9D}:`i'9=0iї5iF̩ϖioD8j}c|+Dpb>ggkG_&A*&_*5K΃_l^5K dػ`o+By#:JLcF*R[ONccϏ]VJd?`zܰt<Ҷ LBz> } z6XI/k0AXj"a11L?iHזi(A6;k쐲 %fyb^B؆3.|}|n|1Ղ2G7t~Dܒ(dazѼON&ׯςK2cCÓۤ2t~f:;oOGt4B(jt}E[듥/2"C HU@j@F f)Q$VkS!ǘKeli'%Om(}rlOn7.cU~t2]r̿W/9 &(R{y"^?^C^]2!H;MnE#~MT-vyr>CB)g}t|NwF=[y (\/}IŘ}1jS1eyE*;Uҫy|l2wߡjul85`1D(R 4b$q"Rc,!JQQƟLONJ=?b쩝d#V Qˣ0Izix(o9)^ q;'y?^0ұL믿G%: 2Hc.EXa*`uk2Nڠvjrͧl>;6uJ@7M-蝦 ⧣NW-i"j ꫲHHPLDY4M8>M|^\{&f z._?EPt9/r" ӣ ԠQ2R,plCvj5iJƘKl7K.':bL9߾`jt'֒v7`w&?~oa%)S;S,MAj(_( @,0{逐/߽!MJ3%x' !pkA*c+5?hz"5n; !iƬ2W0DV*̋*j15e:1>,^pe[g|ΥkVbd~$gy_R;ilo)>3ve;nЇԖ :19~7y5ɫe.T›{$檅szb\)e:u6Q&,df/G&Z8+XYhI;.."-qedq߃{6Y$Jeʁr(v=FpwmgV*w=]κ$LHoE2\Bv TSJ#1)]=NvRwQ[U/ DQ)4b-9S_kh4`Jk##PXkX- Q ,86P~OTENGxμy]v;@$Bد"?G ,[$Xb;ϻX]ؠS$}G*Mn_]HQp݅b3%,KZ/A[;8{22-XыY:Uoףg֫3ۼ| t#~QtaTNx=r|2[@@D F_c-9v)ߜ4Ԗ( a9'0R(M^3iu>~:cbB+;g G{Y;EN`;%1?tThL"bcbI<71ɖ\sL,2D2k"a !%2vR؈KWM*{ǖݥZKԥUî^`cK8QR#ptDTD C°yabmcoBl FLj)u1Ĺ[X#dבNLL γL&$lk^^ `9 E신FPWF6 [sgaX2H4R2gҲ^zIkS:-a-X솵qXL::2zg%+~xVPYp x帨%13kA]WךּL>Η5%ǹ?1dP w||>Msp}2郹|ylHyzM Ϧ fZB⡐ t,C+S©ejMp," r4gr(JPY\DV:vMc:ɱBPl0*qD5TR!ݧ4Pd$ pR\BkӀ ˬqb]pHZBD-$6wQqlOOY񤝥u*qsӲs )HDͦ㸁X jY L(%9$ّuѐML=)}`IR3H0,TRb!tM.Jce]i%q`ydUQ6v;qQ2ɮ/~OA̲wN`~'lRݚwdh':+炰$㽓i`|b5AwsH"Pz9m']R%*}kAqr0܍9Ik-+n0YT˔n0^>Nr^_Y/#otUS9'f͸>ͯ==~贳LYA7̰]wޅS9Ɂ+3O]7 ~eZ>x탸}[Wurɇ~K9* \uf1Ysx+[w8G+YvбUL>@ G%CeKL_b S藘S4M& @#B)c*5S."&0IB5*(8:FJ1b8FK5ypE.g~G2'gX5X\Xm{?lUմ溚NvQ%󗊤SEH ]52!@*ƈWk2\]!uY6" n5Hʢ`Cb% 1tCJbl8F Mgoio1qm}?ln1R-RnOL8FxS-`"!BF!է'EGN?l ]O$qz`q5 cgN=Q"W@Y5I8n>N08eRf$JOouQ1*p C X`th~yחzw܎R+;-ˉJ0ٻF#W |ixJ `ힱwxF#n)R&>lȪ"Yb$Z-2++32/"3#ҹcsuT:a%oUvrȣ\ڹ>GfՎ6coZm Ҽ>kAO|b^HWimپZ6kWu S!k:E5-wa^}=F“&6U9("n8lJ^%K}bƝW?C{.}B[_^U {cp9KϊB" 4_4rAeG&X1Kd)Z:IJ0Yvv GZPE$(-Uew]&K%)DΧ_2<{uUf֐߂~74^45m 7K8_qt#tf_e7 %~?2,ViٱI$B>u)Km65St^v:;tKkFB5+j|hbx'/>ZZ~:MR{ޱ!N/רQCtz֎.=¾jEte@.Gw-S~E]oӻ@::ףF+ v}\!B%_aŮROd߆jqf]`1ҳlH0ѵJփħ j޶ix˄8ȍqVIt౿cՈviKR-A2>HzLjɝq}V6>Y6jHQ"s1A=}ٹ>Wt{$Oָ^"@KM rH# be}0z)#"b1h#2&"Z}~:X~{Dsh֬.꬗zv>7e_j2}[mŝ%T%k^ꙇ0Z>\8-nxpU³5;F2t9֢Tɫp,(t:u"Gw{eGw"pH洏p9M$YkJ#2bsE2m(A=Gd AFBYz` ב HRtL*"Q* J3K`@p0l-,jsaW1,sk0ȁ%ɛg@g)\ [Q#1bJrѬ\tE'&ED,_=SUtQ.K&frjh EFը FgFij%of4?q2eЛl8=3 ߁&o,F z7[9eu/EzJoƓOyʕCDZW0'x нQ8L#p.2^(ƿM cP$2LO?3p[G\d&{k܇Sp%+\Lq1o/}¡A-Jjo);o9[YY -oWT}&엛77ir_be fAbLAϿ(_дoO5%>{0 `/Q6⯥B@5!mu_d6(^J$kVL"t2[o{7kceEڋ<|(v^`0q̕SiNQc4BcVG0vT\aT=6-/ܴ,J#w;t#ĽEJFCj&vSv^SuyնWzV[{nL:I;;Z˲y%A`߅yk'j:RjsŸkc*:rl[juy|1yaz>lJa^i%R9HysQ pʞ| bX] `yDksq>dKJFp~mR`2>df>ʲFvxҗkxHcu;vm/їv_Ѵݫ+t*mah @h7 -/o~sev9 0$ E@J!۠L3Ń`C ~|sLzƀgys$_ZƝ H퀗Ʒg[{W:oM; .=T $_Ymze[@j!rF @jRbr𿠹ׁ =M /1K@jXH>E"k/%%\2NYYh6HL`xqmq7Y,./M*vۍ`ȇzgʆ\xx9^L zPԛ9>֫CnWLcZ7ܖ+v>.s8W% LMr͋'A  Ȧ3{9%v,u*gDEOU)߽d^0{|>{ #W,XPftTPjEӅ픈|,-qǷ9t-><(Q@_$ybit%TuP i>Yv$ ҒZtE9 )A@ ^]zmLc4 PQhdE,6aTT pL!0̌|^*\J T`̀P}&彅ЪVn~ir&[j"dj Y|?8[a7ޠ|M' 2T nJWO?z +W꬚MY9"Xe~Lo& `zb~j݇A?|7DT<[T DY߾h]tfocߘ`=eGt*tj0F<0#DNp4"IN+m<~=p/p<R ʙWȢrLPiVX<( uPG8y{{ft:wDRJ&,͎mr˽ƒP B-v>2$J0*q9s:R$Tc53,H% fKΨT`}N<8"*#)JTP ~$'m+e~"܀ј ? Kr_+L7u لtG= 2%/}sL๋gqDhR*.pW )22f̀/Tnn^]ōSe}n%0 e8hP򠯐\.yൃ3&MD"> o#o$|*DqAihw= sSidϭ&MU9*΄x8#kOwd_ mf%3VFEKK3`@Y$S9d^.)H(%ۈ%bw#G_)g00voq|H.m~:˒#3믚lAI(K8kEv*YxuaO0NPշ2e~p=>8zNp~7yhKLrC.} ZA:ET z )eU 'A$QT :7:]6ͦo׈]6>_6$[s==V3́H^;aʓ)8,cI&9dHZ#j~D8 ƙ(TP̅-]m 3o?"'2hL82ʜR9 xl %`GSzJ<(?6p!*#0|H547΀XQ@Pπ“у`;b1ې~ԍY]C5 GS8jԯ 6Ne]Fm )))PRS)lPΥPJ9p{ @3ۀ p9Z']4m/L(Hg8 G^k>C= `-N,dh4T [-|φIϚR$d/&En_fJyit2U `e㢔Ǹ} +,W `/?l-=&O45W( {Ćڞp4X: maK'*/xwex`RWֽXƘk18N ނL8& M@l/ S*Gegs+2]W%nGNum;PWrwnrCUNKEgs~g]b+IVXΩUн`h[]|i?2]|dY"F+Vd4Ҷy]ݸ.,-?_x~6M~E\.9t^> <yHkϮy?<3y`NI81?$cpU]* Λ(~`7x{T_(7׋%*v}:[8 M nv |Pje'ce(6 "#H-Gei͉t"8pp,~<]!~5Rλ+7)J!$$^ي^{-THѐOm>H$nHa'^ ]/#[<̔j|Շ/&E;N&qt@f 3|be4&I5A%@XDc^.tC8ʾ# KB*$'gQ0j/ւB{@QYۅmXdɘc8(8#&IEXP,H9L($E=YpO]XKrAI;ZN ."ٳ-]v%3]BQ6꬞هM5~ݣۻov\cMS!rjk%N*EYq'[uL=L-gev8^!)3Rsv2;fpGQu&oU.V7dq,ae u [sѨs?{dGw9Ts=ΐr+hoy!Xqr2 Mu}l%_{0:."EnHxBcp%*ru*X.hFɕ0:z&|BI ,pn ̙s*25ٚ%;ԣsq#} vQB90&asL) QlMS[eouz-l74'Y M&5'ii iՖY&ɾv18HO;"v:ez'U}U;iv'Vlp[K;QZgo2/U2$jM:82"'#V&Do|ty&4J#hwW@8(D@3jǘMU0P邂Wg90{/W__@h2g2Lky'wrkK:?>Ghr?:dv嵬>}2u̙ s*aB>o`0R+8I' H#sI ^ OP{O4S%#U0Tq҄[Q т DD1cPen{ge >$d)6RIx7bc&3iH E3#` q v`8i:Db2!k\)XlS-/yf8[-3qZ?ڋ:3N0_⧒:/1O3N߽= 0ϑb<2C+ȧzhf뺺**uu}3O(/A<&ixE (y 뷗5Fy/fRg|0~*P_fO=}h\qikkdN4X2㳊-]ѿa?^ 맍{oSe=YǥWA\M#5wڗ޲"AӊlM"/Fl2UOC ocI^rMf{gKYP߮eeʊshZ9vMboq&8|[|+U-[\mNT_5gX:6Kz;K wzn}L3i?6xqz3w\+Գ 1|[Fӊn{6(ڞ-Bs%j,[?mh򁿬?Paհw/@x!6wul+~VlglmX;H-)'B@9ZÍ43}dk=xmorCC nkgyA\ mgA(X(vveU҇Z.N6&"5l/dGֶ ey 3$-46_]G_^<hq9]@|4w̨6'æŞ{9J KSP=Sxi!P=j?{La`L8Lx;攘b]qbw_SKPO^%6~_ i28aY|͹/ wyYýP`g‚ZQ|^+/֫\P0&X(-CQ$TE#\`yi v>@7v Pp:A`/ `eH=o{-<{kWbD{Oj^UٹWR:P-搷U~goѯx|[3JUuxO/2N.qd*n:i%,F"5Cbun;֙^+郇DŽ 1a=36p9g'2$9bO1Zks Q%fiuF9l]E׹_ ~ܙMz@'?`"('z$.Wu OSXX8DL!LDO,hn0Ka@ݤ~lr;hiߞ_nޠ~V`NYߪresyIO,HXe ` A=*C0->/$Jso;qIx@ cFRՎk;7~毘ߛ_>xFoP๼ ]`P&7ǶD(ASc,0MDXŠiajD[bg[ъ7O^p;2{e7 )I^#oE<dZ!7`bDƻmaD0DRd!E l]CqEOĉ>#()7D)˸Faq-)t ?OVyqw>si媪{o-ϩmq0fuCtnoノnӊwIpTYEgJZ(|F Ny㭵 2=1 &ک{CDrY#!DFxI.:KfAǶc l:%x+zr<g{=Gï<3ybֽgڣfNRY$3S9ᎈj#Oe5;^˪.cnEqgc0Ul[z>N@VKUQgZ82Ч{cA@v &~\S‡mCÇ4){$8UWUWWg9/$Gbгr8zţQ$RUkoZ{CAt[Un5%\?^T>ÇHJCS}ɪ=0ڧc 9Ś҈GQ`'9L[A,J@:|l" ԹN;gyHcRYVQfi%1`AP }zǂq0Q}{퀓U?IdV8* ca`sQTs!i R`:? Oښ 8`('ǀ-$BJ+ϭ ^AA+QIH~u@jHM1[Q{6 2ʴNDa8B4K ),@ 7|aCZ3ssa/cXY~)Z@_e?o$]'wl.Y,\?qG0` l2VFʔ<e$Ou1;QĂsxcn;}b0H_XPpE(MdA/WW/*z ftLPgzqkH$+,KWޅbzQN6g~$GYRʏH) I4}Vea̪r?羞{5]IߘK)t*c[M)2F" TWQWdAJ_oGw/]edƗ}O H0Qk}1܈'_Ӻ HN𼁪5n8|GrH]k/\}5uawވcxD_sOu&W?%ȃX.ѭ5!l&,SD͡5rl7@*l#boQpV67ƷLI|rWӼ0gBN%ZRxpٻ^流Q6+{ba+,zl\A>Wg@^ekB|a6 ZZak0|K)Qa>xMck 7f[פA1J$k/bsMgr2|?ņ[|{K[U?|8_ þ\=V;5F3-46hut`# 0^W)}O=K 9zPΠ(Mv\i,F{4q(xΝ\|v 6CNouC'Ċ6>ڝxЪN#t ڠWuX[w%o!LqMJ:L/^6LrJHt<)/1_]B0[ϿDT{]0j3} w KShWrqou"MKaDWӶy'n7k0!j9bT Ⱂ!2*fLD[%k#u45HQ9^RvE);-3$RvQ=r H$Î{b\())yI̭F<1N[UVNVh|vJQW{H)A,B0l=2^HOa'(=*+;"Ѡ 1v;XBop9k$7A `)˞A9 ( ݮbe'9/R͞sL9ՃO]kLMups9sd;|YΞ*>L k(_-wSk(y5x:\>. Cdvy:`޳;6ՉՁA5/7W7wjcI[.RVv70D۶!F]|2rôr$;I K0}ȗ|{3IU|3:92MU~82 Q?BQ䫑{fɎ>WM Y"jxdܰpaV~wtm~zLQ)MԤ"msev\ Xtq0ﴫ*6DDI}=Tfo<(Lt'oK|.;u\TL[-1RA9$H"3)' Q RyLdg렵CƧti8v쭄'tsJ!o<0\=ԭroQm:{Yuڢ|{J{B_BOjObD ]sc̱|:oD=<@rFϵ,wJÃ& #] 3#v뾉uJ $y5lJu" 3.ڿx3Wxf7mJ1,,VJ\e@"g7D}W9D'=ҰuOi{}+SRRZ|TN9ťoW12#XP1g&B\;ײw:©9%XCspO/n⪢G'l`yПL;Phd@aR'ZK(Ir yIt{t$LEt!"'U2ɁnYq^Jv( `9AQN^ {P= Ҙ60a+˜V!!(3uS/ H)NP Ij$fXXCZiDڳF@bٳpP?mg]i-(GEEҖ,g*n~nnӁ䋋}Y:1K @OKwA Px2FJuJh*]^((\kDI;%b0ͿxMڸ C``?Rjsfȵ1OY$L0&ȚFFZ|^$n{ݽ]w q}TԚ#[״X? ʕoMε\)+Hf٢\3}}G %5z'G }lafVFgP\ 2H#1DK,W#(<']4"ap) ;I8F-DfA2*K)O)aZF9{/nIWAYk%} 3_:^"gFy-Wӣ)lòhf.|lÜQDr흵!HN4",!F0QI)V*~@%kn0p{t4^D*#R"TD2'TfܥV,|DyE`4}< @e>CkoTF ߢ4Tlno:I]aζJ2. wU/lW,œWY(ظg|΀>ׄlG1jZak0|K)Qa>xMc,7f[פA1J$k/#-[?GZʘx6,'C:Y<諾{L"l|Y  }2ޜ 1iF N{jIXJ輀уr7;t#ĽE8Lw ;i ܁*"m!U+TgUΫ5x`zAI6z Y7g{AN0E-`x܌^UvQ`cvJش_6,XYJdr%\Ј$rm˃w4*Yv `+$;, Qt~OqZїi{ѧ~r>?؋IUBYr<  tn"{r`/jr\*f;8>e1gݗŌU -x{_֎%!Adb eT,T,y?S1>%ZJ~[N}o:0M|d~U X?{WqJĆR틀 0C v&0J(&)e0=dK*n.VrNa!C ῡ/~QTbM\ɬш",p6!NcN_\OX0YVT5?_~r Z2S&4{JS"m:;RuW1P fZHJ*`)`Y5\W3.vKJ:3o[r|&X.6Ϟp[Y'[&5}<$]y9oU: ,A)\E(?f߁^tɲiatIQQNTMQ1+ vs:ަwBʱs=ԱmRDCqx3&t4MeXKKyˀY,Cj.k+*ioie,:uy F2yevX?W g~%owWO:5}4;]i 9A!_Jo 5{0'"M s gI /'0ҵP]xeuҺ'CϨTS}yGk͸#ϭT1WV-P=y>qtp%Y+1 2o FL_OߥCm1M? |y$Pv ',DDꥦma$EV 1l56Sgpk[6 nr;v&ˆh]}{l#;t0+盞oNod =!"}JDX,-,xGI4 GrMR=l2|粽x)0w-'tn3(So^cRǖ y͍iϗ×9s g3sdsh _]zxcؙ扊S Iw^9NR,6i1L6K<4z<(#wf;%)a%6ml4-cHJ"N w * X)ṔШ+#jJCtOm79' mיM)F8mY~SFmݘ1oLk2QuȤˎqªgN^ r}}5+7ס񶭨zހPԴq8f2d~|ypbj $EL9AZ`4 ,AkOY_p̚`"O)N6xa&UWp^-#l:yYu*|wfBsi_˻?iҳclzI|$,&r8fm~n޵mL E&6f\&y˓v߇(ֿ 웺yVڶޛ_wѧaeznFk-&^uHj~AP+FR+jt H)oC@*I(U!_wKRJ&,ǩ֕{A(׃o4[;lu2|k,!U:Dn!Q"WA1 "rtH0 *C0!B*Caj3jX$"-2b=6M @l잮-ܽxV~m]ҽqa9:8ܩ w1aa~Ժ[gֺOjI9g=MI{c @aVzCM{:&܏FzGMiΧ?oڵwzrz<`.J`|#ҡ}֟͏t6׈q೹ *Qǟco}#;$_rCɌQX 3^9?ﻃ&"kbIu\!bb!̯Ӄdad.36W5;fgd"39.f!ye:񚟇np>p5H,J‘͈G9%ۈ%Z!.q(#GM$&H 0E.R 37fo$)Ű[)rcr g?ITdv q̳[JY os=9rL*]?hŝ~&s J3> n^T2*0} J\B:`^el.D!6;ȅ4*Rc  D9>0Gz*198A籞i@ QzHL ",iP%LR#1spk11sZi䴳0EF_0PWŮ;8 evg ?^VRh4E*2xVDLN[V+G+IegE q&XO1 slLvaF`""בJIh7r 8a(rnSF7Ew9lVJH?B"s6P$t&I&w O筱D}}yuYѷOG111Cl|4t0hP& PfYOfE@jkP<4`0oWM% '}]R+R[2EBLyL0&˖w|z7 :^W/N ӓ]"|b|D(؝)^V 5bHK.3n2*F6[P$\!5eLi=N\0J&'al&ICX-dM|#{U3u}{ͺs{CR[=BM.PGz~z+A_E_?4i6 bfL@hV|1TQXW?rA/+ )MdEMWŊV9~ϵjخ6BvhRPQ0nل(.K[l|2:U ΧӨKyJ!MERJ\{ToRjrHbDҸdXq"b!,D(siJ:U0bQG"`"RSFDD b )/V 1Vl2};I~wjĩַzym:UrooLJ/5@VGi0N"e,4HCA! "Q24* .QAsglrD`@*:d2sp8qBձŰ|vw2l?SԾ |$3rŧax4U.&PqeѬ6x`t,,8ZI}^Z)Taiݮ־sO`t}MxfO>L_O.Is]<Ӽ|`n;/wgpkwmvrilGƉ+٤ЃGvz)5NaVdY7=ߜ$`"{&CD1X{))Jnpb΂wDðq$$3&w.ۋ ]~s7jQSw-ZbEI vLj2c9ϫ7t=gn̞1{<cn{m#3VoKPOhy]1tWC$UoPRa?y̤msȸM]p/NIm{I˛-P'ͿVtCۨ+=[g.?~ `gnPCj3,P2R%/:66=&iM&MKR.M;yҊNȁNzRլ\۶IzʻBQp̛p6 ii%F*(gSId3DV)23u!M|0stsLpNa0>hKM'/NϱqRH.O]{W# )-6?(̥vWr~=|4<0Km*CA I1}_M"_D_= TkGT Ⱂ!2*" LD[%ˊP0vY4a#AY@SJx9@U63~68Q#IŠPjXJQwnsoԽ{c$|2UDX"lHXMiV͵NjC`W^!Lj3aC^bRb^`tE!DQ9|$F`p/~Z>Nn39WJ`hi)1H-h{P؃`" ;ZpRF3&1J{7e1LIvf!x$o+7{V]x}40h40A/I&eHNO2[dG31G^5ts\GV"owu,%{ߏWCu%|  g%f[aA7,[]x4!ƻe/$ۢYKc6_;oqu Rd廳u %uq‹Okʏ{IӴ^zq^]i{F`į&◊xBW;m*yF"pzB~FWPno=-sٲCzo_7Y.f[֘8^qd%jc?_ ?^U:wo}@ ryq֞!WckjMfVV0zP9%KOsK3:vVUM2nN@0Vg(gH کVfi U؉ ʇ\[4#a.9<>nV!+y #axhG*hTV}T("]0P#71͑0|H8|v`L:I@ʹiq}n9?-a&f)HS%,+2POKh2i*LZb2᪾ś)!ax:^Rm`})ȱIzϑ0|Wr<dI&BrqҖקV6P)D} C7NV}=ZAOt NPvNw.-;{5ixJoF{oKORᗣcZ% 7$3ٱ4ۻX'ֺH>>9ɜ8 2{ d\?-axt4W$*=Tۜ 㩶Q8zK\,ޝ_\ Pgnow*D+}K^@i'b?ξ;c3N4tp>ɬS G&rl&Ks$ ތ_eK17?g(d%,Zb4<2GIc$,Y:dE 2̞*XFwF=ZLO (HL*y"QZZqs$ u LRBdܨJ+I\S Q2as$ _{)I+ 犋1L!axѻi$ߑk/!,WOǚ̈́ϑ0|0z}tQUwL09 ϐ0<*2)` sOx,d9Nϑ0|ELHGJx4S0s$ᐋl9YE Z LTm*_je}a֪rFrT9Ȃe*qd!ьΚbϪz'rMՈ8M#a`>X igX*xh909]!XҞ$t'slX?ArHަ^h&(aP3( abJ9Fw.k0Dw[3*mW-!ĦzH~|DmhLyn=0..1#l- u?C^~0D֬@uRbi25jLC['9~kU ņGčOs$ BmryGlx!W?\c>W[p.&8C7f<%-j0gp54!%7g[(JB)}I'Nk-,3,L@5i-QbV 5&tCd_89m]ry{&--Kn*FK otBfN|ы˵m{tF>ڴM5:uA}3sQ!rsVVWx޼S_uow>ֽw$yyD+;o>yWOn~/۴0ϭ kOWzuOR+W>.|^U9e~zD>w+yEڶޡ'FnɯI=žcs*½(l? u?)_{r7GEy%Cc>GfyAgVG.NX혭/sPp{Y ZMF[gda͇H7={Wl!~kڔ̵bqiLFvgAhcG{A6h"Aq%mݐz`j0ՔY Q4{6Ҏv}ZA[b`A,PXVc{dgJpySP^$u'~5V׉b_3r^,X~%!,} -eY"- `?*}C*w'K]-MOѻ'` >q/xp'EI7~gW۞w}c_z{)Z݂3f#IL'`N^T>YZ›z>#8V9V2E@)gxg Eѻt[:W $X!c{HW!WX~ 0{,[0)LZ _̐"%IIMG$kSS_UT)ț$4.r%doFB#nqqLVݰa5㸎K2@?st}:=%E{Wyy?f6 rODXPrp s4 dI_ۧW3M,y,ʲIi snGI4o1 v^&=YǏkߎlP N[ϠOEFVכ 2jfN ~i/ d> M;!j6@M'K{iT&pb 'r_ܛbt1@ xݽ9'G2+]>e?˱*p?F;HPLMXg8ޑM&xMVWJ4Y}FSNECޙus^\c%2"(0+d ^*ͯ B$+ WΤϽH9B_],2FA#K&7YB GXL 1' :8er#^P4 y4.jg" 6A p`ۗ.xVp.rF8)٣}Y3{+Cbr6Q|P!wU6$PDqˌFY8 ΘIgy@FDhKn1ɭLN{9;:MIoY=Bem=AghQ)Ư$5UJffc(2LQ3GE2 k hp4̂-Y Tš{Fv*yU>7d,qc HU2OJ=UI+B^zЪČ`5Lnhw\rhy8IGZi"E+Ǚ$%LG=8:.8=gzԋs˥Z 8|JPT8CDA # qC 4&( gRXf,Yֳ9 =?o9ܝqvO1K*IIS>J%4i۷> @Gw,gq-5Z?Fkٚ:+}6g: n0?z{Y1_hĀ9s!MڅVB@GL,h7P!~z$`1#m䮀g:5y9?3ar~"8]G,Z8b4m"<[xQ/!5%sd gKh\Yzxf SKkJlJw Z6nq`^YږqYAGM8)i;!^mQo;0BQSLS۩kafȅ&5[[(jI}v6s}tV6vIa~rjev7̫6 DŽ1g6ѳzlG -q{zNYb3%IVΚl6%5w''b߀\vA8(Q{KbXK.Y2h(%pDc{f$}Xr0dL<EWZ ﭦTGTR04} YrXeO !˜`ҷJ|z廟rX$ T ,,j@R( :jxI7)eTh]ٷzdr]*5~?ֿT9;/Ն&4xNԞ[77DH}E<Ȭ#\8`4E /+5W9XEt}Wq^F;y \8scc IǛT&HgSG'o"<~{wot֛nu|vOz.} 65zfQ\䴴oCmZSH.'yQ!iWE OL3Ub"'2~rA`[z hF J=>XD; D锨ăCx =|"W&=C~k~2j۴˶jޑj%ߦ9cc&2BYuh;[eKF +-|nYS\.2$'Ӑ :f7woo!Gbg?$|bC\!q}wOv5EDb<ǫmMٻ8n$Wt{" 899\ @ cAɵ2;g4jI;eW DhXz6IDW)M0#$x x!.-3lyBZO:~}GCgxC\[_h89]9wg,=췶VFkwLN\]A;ppV_Zϥ)}2য়ovģ'Ľ"F(u 㚪za `^t0f\p<4 ~Qf5}s\`|s9/l}źkR&@ 0؆V $+hʬ$g I:IWޮ{VV tш+֞3'J/R_}xngp,͹&9Ւ$!lDSCv/rLVIXmJYl ֩C>jQb.턲sQTjKi^?0l9"*;;ZkMoDqsZ;RZk;1jLWΖTSԚT7+l $e7ViRň9Tb6Ʀf`fEFgR<cȐ!6) w0ᖈfzokeY<ZǬH{wHy{c5҈&UB%%sn*d[b1N, !}tB$b&u4bIUZG[DRոFNz$Y2d]DrU {2>U *^܌S%Q[+DE)$ F[4h!ɵ(sRBRLV̇mnI2UmRR se_'Z)|0 D6}s+\&d5#(xKd2 Yk=V6z\. DQ̤OI=k]$Mŧ >4J4)*li_"IF>{(.Ic 0gM(TGhOU7_*KɲnGdjɣ^f D *)c5.d95>fqh%dl6d|nlOzYvLk4ZI03JxfAmJM)IQrM h ߬t Hj/ֵ:h٫)0LCb{i6َ-$:PE' U4ESؽ.Y;Af29wykM.{.tbs4fLi!j kc{Lp 1=&c{Lp 1=&c{Lp 1=&c{Lp 1=&c{Lp 1=&c{Lp 1=&?!~H=&<0׊cֺڎ*@ 'Y$e__'ݺmzX/K?zկ"Sì;c|gz%/~K_bX%/~K_bX%/~K_bX%/~K_bX%/~K_bX%/~K_bҗ_uF. v`WFkK2:V}gG]=%lWuLabX:&1uLcbX:&1uLcbX:&1uLcbX:&1uLcbX:&1uLcbX:&1}:&Tȏpь:x#sSTZ|0J'}:&<'1]~a'29#XR&2L,eb)KXR&2L,eb)KXR&2L,eb)KXR&2L,eb)KXR&2L,eb)KXR&2L_ݭz$˾մon 7ZN6:[>)!)`p(>U}ʫnePWJGyc%eWc4kjH$=Y;/W~tf!M~skݚOOm^sz*0 1`<~pH6cցUpN_ [@Gw BO(:c߿{8玔lQh3H?\nɾۧW),>z,He=cjx!ۼ*q^库-ώ[VdeWT?ϦDp2V?\Q}7QV`drD{X9Xjtrr:\<]xx|ϼʳm>|ɓ?_tI{ooևSJՒt1TXAٻ6$UKcom^F?m)RPr~C/Cǀ-k쮩uu&YK.{!m421lH~Ad-KbZoP4j+奜źDm*'d.?3-cV_a]_>h^o-`W'8H0K#C-w91iK5<[Ǐw}ɻ-Ӳңw囦]:gcNYr 1D$_43OG;-qqJOO&iVCğF4*qXg e+G5äQ̶V,phGI/G8$s~1F'Tux"Io*J純I,[?zEV*_=Mc7s`"gR9:<8OeCa$f1=X2*Bem߇yuiK&e0n5b{|6½'psBFCi| b-BYyCdn)l0UA?=UB6"}6Yc32O@/8,gVܹEA,(Z oCE1+%;r+>G=(ea>]8Ftc@ܙi{hLG3sij{k e6h+QN7T.o#4B7}\xE6 (דڻ>bT\ Эmح%pnȭ` %[)j,@sԆKǣOxK|G_\pfί*@Ͽͨ3`i >xf'^a[j5We@%45[J4mm}{T>&_ʾp:/GZژ f\3s]Upmv{6CW99>zҠO/Ut#}RECO ZIXfaH1U623XG9D%,O dQj,oC(aUg>"g6N:"a}f| <LWKpQVl#S$"e1kX-[ĥw? ‰D bEB….),XEtsHBItg]TP:#@=;M!lmd0rI1;f5gH­%g54)V`nN tEC}~dg>Y 3~*Hd KX))d 1}VgViqo罷wR5+BAT9J&!8Zo7L&vL{k-.(QJ$Qcj~8lӪmm!"@fm]u&"|DqQx B:B xeͱxWzM8gWcar\3 yv| M kp ?oUi*G }n#1![n7WϣfcJ=Ry]i0D\0Yq"&97B<}T*-|f7Rran/w,y[Ŝ,qލA?¥3xJpO|}MH"v}r> ar3Uc[>O *`m`rlNǼۦ?'[P }ӑbw#p Hs_Alm)$ s)CnfQUܒT>(8u)0>Օ_j]1p@d>W*UKm*Vsԧ?_ ΡȊп|D`™, kċAE2V@4K4Se,F6@\KϣK)t7GI_#Xc^k Qs_'jӲ-^? kv䱨ź,gWs `/c7/VE&A|f=uFZ&l~3X| >VXC}?.ɠB*]ک/_  `y[D|"4@8)Q`"u,*# A*Ͻg&e^Kibx:ʲfP]%]_B'q^S֜cԡ@H`k힂mY5f/xfJOx8?~^)K!f. Qs E dez\+9fg-hF)Qr,DABh*%2hhTH)E\ND1BfZ*{3%sӞJb,;A]`^fz@v3GY`jD& ZڧW*ه?dW8): 3r9b`u7qu(o=y rd)|.zŋ]A7=fH #-0k.h(6)X"lFQNH8=x]%ƒwa 㝝,'& 8Dcp4`fO AR,e0rb@k x`1*)IIa:hy}'~UXA<"iyuaT&\ԻP@Ja:>7aI^3zhWƦ'3HlfG[X*L˖zYɐd+Ⱥw@Պ|[Pp[Zbw7 h"hѕl=]7]OgU[/&2:)+:B?%CFow^< [u˿;iV{yv{lEV0Lv7tr\#d z68t5y;]k1'ן7_旾w(|O'}&L{b^mc0hoGn}lelnKBıdž\Ї}Ⱦob8M rZB;)H$DmX/<O:IHLdJ*a9SV03QYC$ ,(dIwށ*X6$ &J ! zf&&t}jWWnZ]}Kt^;d'Cwvg}Z꺴ۈb!F& 9]V-2JQ%* '9 ZFet1{Ot9e`zR0H,MUQ*ˠ][ow{)b ;Bձ-T-ŸdC@ Υ0,viiGma;0рGpdq/"pM0vw0YdXžO\M~.4 'e#YxJtM uC6h"G.]B4s 8N"!{-&^Ղ0t42"t6 :>1m6#5sphT9PWقsYKA^_;Z㦜9E`=垘%7 +[q4auE\8:O>P 긲-t>EIG,Nd3,uf.9E!~p~l Z.?֯gc[k^\)"!qjT4DC K.{ѯ*JѪb։)@H*ۧ,W Ni`I!Jz!T1qU k FYWgu Hfjݜ]ޮWlZ.m^p0:?]3,KOtW8 lBn Ϲ2(5 F)!ki)G9$]H\M9uFӈ7:q" FelAPJ0BB 0Rq* 0h,S\Gaݷn,m4sXƃЎΘv8X*X@0LUʀJS0tTSVd9KZ"ړIZ1-d0bVQRGk`lR]CzH!F* stya8y- 3iD(2E\ m875(%4_å5y"dFKZ ?/~NIƩw^[bjpٛy^noy5a?ߦ6aaҮi] ,?ΟMK"dsճ]p{59U26k(e-u s/-d.Y YcmB[zo&ʮv 9;OWߗBo hF)=oZ8w蕶&ٞ=jقC:) eeJi?ڡ+$_&H.O#|b5[Jmew5:#b(RiJI~LEGcE96&X{3}vG?Ư̿ןLoXr!9z`FLpbz3şo]ƽf7؅~~uӑ/_ޡ°Pǯh.nC nEy4nbӡn+e~QØyAht]p56ٷ69W&sk ]6 t֬aJk[%gI >H3mT^^|퍆7ggoJͺ96jv,>؅dgӾ'.lYQoQ.爿~?G^2UKlĴw__[ޠy7_hY?ƿX'G1Klհuk?T\K޲Ͽ߉dvςEmc8I.D4V840Ǹr@@,Xۑ^=[Mv+(OVf *9,*z RT0m!{m3e'mRU٤˫C\M'Xy1F}n=fJPz%tG>*(c!̊kJWJq'E 2 NRL]J:+{v1G#/'i$,15w+ZHTHp*CT1M px/)Y Q;Έ |Џq6I@ni#PQ5qsŇ $ hb*OW?LZ|(%㍋8#Z$#qH7rND4ld$rY.AT3NA&HkJ"<!)i9+dtZsHR$ڙ %˧Ĭ4kCyE Jjs ZZ$QΪܪClRAQs!rU2#p'4aG3 4Zg !4^nI 4Ez ?g>(x s$?VEEIMK&lH4[%U1(|PwP Cڴ'fY)SD( *YRF24ʙe~[GqPetlxiC.xTES+Bh'fkY2FE)Z廵̊+"([ Hzr(WMEaqH6'Ȑ4ؠ5#3R)ĹE^'EHvQ<`CjP박~W278 l294@`=J9\eE`Î&WYcU2=WYR.Ww ^ݕ}w>&a͡<]/x94)Bx"R.*N\Ki|Ϫ"Ϳu?+ &2DcYdd<,EDr<31qot@Zt=*utOq.L9WL114ҚE_\ 7~Ic"CJnXd@`ŭ'yNhƿ^eW@JtFKFeQ7ȁll41~y~hMN2AET ÷;RɒS~2-vMF4[Nk7vXA* QS$r$&3\q$&=ۊu+kut h%-P!:eTGteLh93%VeZ kDv[z4ߋ{,}g7l ]_ڿ'em:Y~(xn[\bޞ\o7~F?M=ZO_P6ltͶrh @2dDHIDqIR &KqmtJͳ^NL?.#zRhb~Ik^NfUKD3HWp!i ϶eӭစ[C)*4q Bڷ>jztt^mhS9eTb 4])(N gneTR~|OTRu8Hτ}r''/Oz? Kz?SS^ڭReu_zҽYOr흣z/lg8)f$뤜%n A(NKB|Gyx#|Rnmٕwy;]9݆4sfl.m+Z\\ZE25|Wz f_c2K5l'|)DX{.qK$`$ZT."kqcӸqUH)$A^m޹qycPc@c7 x{].ygw{ƝFJo8]o./Qy1M-ܦ7^{Hckqfn"uj^TtJz\C;ͪ z}ur{BՆ/T_j8yp®zKgJ8_ zoj)ʿWv@Cn!pegJ *AȻ *85]v]7vо1l^M6ojLVWzg/zjsUK7v>xe[ 1mQKH0+u][eL%]_*^Go(djۼenzne(0Ps4gC lhv!- -RJӝ φJ lql7RbÇ QjYgJE/LEJfnRMi=!n_5"ՈPjD< K> )"R"@S&RILyVe&25ZTݔ;;Qgq҄,y:e?}dVIqhz9.~muWw|tt{ DC]xS iGxHCC8i%2U1䈥J,*Ş;~$ 0*dƅRJ2cւl_!\ 5`ah"eR7H⥤HJ%FM2wDF Py@ J*Ϊ~F,B0KQBXYZQx)#"b1h#2&"YΞmָ( oe-`~37|+-voئnDVGi00 `-j uc(dz(:MyU8y-X``F͝QFG֕5@Tѡe~7YS)vY[(ey˒SҬz &,Z9zSƂ`Z0yH'*׻o6;9krCo V<;v,,_T6xk SPڑ1,A!)>|= :_vUYŮց?`͉W!9 "xC=s GK$Ƒ`%+ z#>8K+?bYǖcY<=Q4ڧl 9Ś҈GQ`'9L[A,J@4* ,=\dyʾuzREn^u4K_z)?~{RLw Kf57=i^s9CD?qrRnyȥ$8L!zˊ8%hFԾu${DЇ&D}o0=ZAξ]eL1uVmpm@,-JN):D폷-$%V p.^za/im^9 h3ۯ毂Ӡ3i7ԧ2U~Qm2o7KDZp{F7/w>q򁿮>P"ِU8\d8҂ʀ]{S^QI,A \%\ȵJRz%w%>`I,(Kbr|Zϔ*L 0\hX6 qAúK{L]$6$OI6;g"QjN `D6jH(0 5D+C-R!Bց0LVXC=6+Qn6-ҝ.7yٍaޱuDLHyQmQ*A7) L(e3P"R +&-Ȉ& ӈ H!S,8{%cs ǗkOx*D `}4jءBKFQ霏)I 2t$ ׌z>}f'QOIp0~Sak`Ljʕ (*θC>*OtV6UMc֝t^cQe tEW )2]' 02`!H:b80gÉwoW>bj1 L4AzՠrEHiPrS`V_ ceb23zDR/\cŁya9J69skԁvuTZCѨ#|1W 7z|Ȇ!RL#:A&I:3`HQ)!!_T$R,!ˍ;On ,Edk{>e*]_| %gŵ:μcvcUrh[c!fH RM}-SNTknohJVGXR0Ty[-2]*[xzIxҡ߿ao 0RK[6l3 EJossGWxK n>Gy.VïjdgPP̪skQ6ڪؖݝuM!ABHKr5 _-~ Ht>-1otMm%QpeMcP`*U0xa2،)$Ť3_c6zu ze;֜#)Xka%#B+eT!`if|„HUҸ>Yegg!ր ޜ:e˫y̛[5Pww.oi-);cLƜĠ;̵Mr񥷌_(kuZ8!~-]Gjwx˂vHrrKXd WiKxyl0D(ȕBX`ha)1HWwJaW) C$va}l-PISw #[x ^a/cfU d =!"}1E4^JZP"!&`b΂wDðq$R J!DûPϛ8V5zҜ9/@Ͽ>u ]h<)4@EPFqKIr^y3 #,[kĐWSK-y!;R:fVݒdmo;&A9z"▱z.; 2,FYGs0EVXM^?&GxU\HC}du\O2[O3 .f`‚^D‰uXԫWPF۞W'su_W~pG9HpS";k)EВ)E 0_uȚO{IijݫfO.((EcfqƉ`)XoPAc.b#pTFw.%URSFDD b ɧG!eLD,v+uX7ʗ o]MϠo{&nm xO nf) I KqۤVM6Ճgg^ :"daCR@wJ xNJK<=}ۣ..My`iN DiQ?mK]]wG(#̎0G+uG-tusf⍉b1vG˸,ZPz_C$Q Djnowd&i%m80%qx6m}fq$%4yLÃOÔec&N B->!-X}v6syV:Cq&2xbܰlP=j7z:vLc~[Ϙ2;&I9EA3q\?`ɫ*5u"vi_j_*kfYdwwEM7$oe7шi%F*(gSId3$jU#pne0E/F68by3-G\˨6]<:k!![ҵ6ޙn1J0 `Ʈdfk)Kf&(UW2#̔B*MK۫o sI1>*Tq%z(dXH8kq u +"J; <JqviTxR2a9?zu(kDBNY׫%_~Usئ@c%b&eHaUPe# "gNG PR1vgF/d&#QHYu2LwZ DEFL&ZOHKDE6pVYwQos:~)5ĖFUڍn٩>cI=URei9gR%k1cZ ?qZ2>ܙȴVЪ]h";~&?bҝ1m\Kڹim̋9>g37FnLJhe٧jGø_ꎉ je6V'5g˟ ݷr{tlX{ n]ȭ|޴:L9M6: &:Z-%0AtY5mfƕh|fPLGoC#$D3VFE^Mq:ca2]񸈚AS 鄽)ݗ<_]A&pVPuօ=9_LdNܧ-7aNc~%R9\0$YI,bĀI9%ۈ%Z! .q0#GM IM ]Q{.0 t# roD1 $ek$rǫK OFMa몟+NubdE 7萐& odk!F"0)y9P2zEWPGm#Gȧ/Eh 2;{ }0dt(G3ܟ$ےmٲݲ$ݭjY|JɄƥJ!',JBa1[3Fn/e'7x.=BtV.W˚~.NNi_Ѕpp0={;,z e1, pJ$ i>i21䢱EB[.P㿐1*֑h!9+ Em\9YcF;Gi)w=_?^T݇!G&lj3Y嘊Rǧ:/\ի SU7VzU-d- E 9jϽݕtOkC(ITjmQi I 7Jdu1lwXaE.w MuSdVϝPH#ot wwb=5-(-=WeCljClH !]6ˆt\'Ti0iTr F֊}GӘQ4E4 oՀy2Dm 6`ۀy0om 6`~S~ Y.oH-doHKH ײZ6\ˆkp-eõl ײ(4\ˆkp-eӅ ײZ6\ˆkp-íg~ ;w{jDY#bӈM#6҆B$IZC!m( /$+$[($ks/t tl0pv{^z*Hi 혜l#3# xwht߀j$p*#1TF.n1TD Q)q0T0T|Py#*C%}J o;;"v1DՉߑ-Kůb={߬ J#$ rHZn2LZN$ELF:{Y`^@ WŻ)swOy?F Qbo뽽Wµۍ`v)Y19QZlDB`|w>>MP1ۺ,t@wDevZ] nkp6JgP^bq~֭Su7?O^;s6dO23>U8Z(*,JrςVEQ]\L$%^ ϧJB(i^6r6+$PB զұD6U㌡*^幷;6M~ +ZX7aʴu&GH 6.!  WɡNuD۟4m$HyeZ&',[IFSdbif 9̪5AW)lK "Df:mσ<" .)g>l74*ahPj.u?*Kbg0oq|}-D_ChOo笾?joLby֗2X&9,fѿKg$_5QVM)b%g|&㮯11Vϫ^_\5Wh܋CRq*EוY ,U].WNxȻ/~OzwG_LOUףwݺ+8 gm.kJ|_O&r9v kwG/)4}2s.#mzgz,]۞q}8}{%[ooh@'ܒf`zHW:f|o|>ZCznKw{:a5roѓ^DЗA]`m]d u2HK8F@AI-{`4L| 3T ͒ I2E)IQx .kТ+A%cE\N>S)fzY}O<#rAZ%ҵ|TkݬT}S N;tuc$L[խ#{Gl{#~q5G2> iXQ( F)>> ~9dd` Yc}+6  9sUQ0E/=ۗAk{t,EOig7FAɇbs!XZdHR+r6X9"1&{5wIpAb|fbFрp.> R!,PB-it^,SZZs.**B& mT1K-Z@kc"xݳFΆrwYNJEF.%Lm D"TigZ$k6vDԪ'i4-.7S>_( _EIL?hMR `Em$cT4hEAD*t$TszslO6I{H6IȄ5YѨ5S, )GhBIg}[iűWe Y,[,{轗0%dL6ѳ?j6"9B6x1UG)f-ceajE}x"|Ⱦ &Qg_2;>AHVDe-90&Fr>+i&/x'XGӯ"qc{jX) _%ݕ%nqK(@9 Ġy@Xe`g;NwXe R}*Tj۰ʾA2vƍ??g3b> =Y :E Zy@I! ֤mKqo cU? ً=9Z8X(`]49F1ţiЫ$[&WNւFqs\Ę2zr^E*oU9$0"m~6_@(Cfw]}>A=`ކ⺻ӏi]*kJ3? }AE ̒.krf/F:(RJrIڴPyszʚXJtQȉ*d ՇT亀o1ٺ)zy~} owmI_!.c!A)$ Iٱ_P"e$6ERjp믺M Xl %£M55y΢O>vyԭD. ҦǪfݽY@<mu :Lf mg/eVg hז7\YBX!4.j8) *NY_{H'5R~YOp嵥X8$j \KV*iE#,2ʄJ8/"ur)D 1PKwQ ygLDI`S̺%+wo" Q 2]\,eO|^˥E$ܬl>LXJ2e:qr@8;Zx)gK?m> @pF(Z8l#)a B%GEp>gz%NHZ @e>J4(8C ŠiZV-S|zl7Vܟ9̈́]}UWrXQFx)uqi*(X '_[5vo qhi<ի]rȿD8PY`'S Ņ}WZ*W*KԱׅRrSBºP 9 T \eqUVc,% WD:%v?,8Ҫ,^#\)}9v~Nj`q5W=s~[Ƿ{Btb}8~N ƣiwjDo?߁҆8~-T&p'?IsB0sOT Lgi8vR*_#L#FNR \eq \eiѯe)M WQPAg˗q4?W]7hW[d7KSIYvj՚1N8AlZzLY&QDHm`5; !9uQKߒqwxvڲ4FX41&lx3G I"h%bJGmRz V>a$S}F8~56tbμ&MpujU/ti!A 54|Q=jGP;2F|J*95ߣ{|Q=jG/aJukcX6֭ukcX6 BkcXM-s_ƺDm[VìYƺ(jcX6֭u+ƺnm[ƺnm[VUƺnm[ƺnm[PYJ*\k6nj+__}?I_@Z]lF[^kӧcrxzQBbhJ.x뜡BDʐ)P5pY ^;'$_G1"zxIaN4Yh]A\"1˼DM-[JeFWxk#dB(i-tRDJDŸ-dSh;I7_i.'4/7UPo^ͿvߩZzɍaZ~2S˳?D"F4'8tqE^(pk-$OIGHH-V QXP"*#,N:$NBOkP!ǂA$AR- lB3,ٌYAf< ea, r05XVv۽YNL|_mqw'4zr2}mZImhf&@ƣtX$!M3{J] qA@&*"6< wqaKpRLP =MbDad1ta.&fvas.iSQ[FmUQ2u/=E3'snK>7msqF嵖<`:_HlAȍ@2\QyT5!/Ydڣ)2( uTbl֩}d`<D,OED]uEĊ4~fEhJkIRQ`) keTGO ŭ3EaD,&U. qZg1-y*.¸h*.V\g۵q[>Wux5}XϲsY䢞=nœ[{x`z<>mUNXX64њFZhJ+:R qGm̫ȫh=Ҵʐ|;ݛ?_n%~q4]FhW?qI\F3?|TWMJ݀8Eh%d.8EybA;V-xsޑ7?h) ~?FEb&Uxl.vS n g^8Zl?u~q۽ f/jvtmwsO? b\chxnN2WInqANܕqy<ϧzM$wAxKvJd5uy"6y !D٥UuW8nߨبc΋l,A$/x 7)e}?#l1%XG45O._\ѽkEjg-]egwMwvە^9T8]7\gW;s>G5כ\-t~/=|}瞺Gݠǟ⋅W۶&n$ȰR⁖ Ydl4Ȥ8.q܈!|(lּY𧓞8a^\ٓg;\pm$%ѧ$Qi1%AI8EoT) >zJM`=76Jʘ$L"I.2#:j>1P<ǔT)}B/ n#}7)krŘ}`>0 F}p:,(/?h N"\ɋi)\䪉TFxb4AhcNE]ˈL[XVN!ZF)=,()5!;Dh)Q- 1Zjڎj+5ĭixE_sw8{|7W7 Ǐ%^X-fJP= zTZ~+ '],;_"]6*FD ciNh[y\BQ(Gb \osRPޥ\d?P maR]+, KEANKkJ2n>&*}"@)˲|N1}~Pq3*L heM@u.f@xTϣ(JTYrPZAbxB\hm!$iH4jɸt(d9\{$bTmA.Lr y:鉶6S~g 豳_^ęf|n=l \6B>4v;g. 9C)voTYEgJZʜn#z&7[k9N<%Zt[BrY#silAX xI.Rg:c%hTɓЋق`9: }Uq]r]WT@ bM{pv18ۀ >i,bGW][*is"ړwzc#Z$(c@b,|dDE341c4t  IFLH+JtfԈN Np("vC!20 }(M=ҽ3|Gěz;WC#Kr(J*R.Sp-dCȿ]x- W8]rtUm0&mϨSy";ƬH+.̑Lb 4l9$x9]};ɑui+rn4dȔQ$kFS湴T1JEs^[crwShD'_jsMM[iZsОT@ѶГthYpUYN^ȒRYCzJWL+Lɱ)`@1Ϙ#ƲDY(+ؾ/ٚD\t/Wd6\HoT'NyqugDdy%dog Rwʻ? h1vzxq.cv%VI$dGIĝDG{yxj222k(ypst7r'g{8W!W~Nv\.0~<ѤBR~C/MƀlkWUU]T\^oLmJsr?:WbGUaG حg/} C`r_Gi,)_B12KǽQ5ոٸ8}\O_EW~9u//gEKD2.Lo6n "l4\ ^:wGo >j%bԽ$,F D=z73Lο!iޫ x@5||zG=sHhJ cc~5ָO;O);/9l<΋p`k0̕SiNQc4BcVG>%9 QqRܫ'~E)> 3ā$ ~zaǕbH`pϹd!;pBE ݤʲI5 S)瞍'v, B!_)0k]c'K.Ux,0||+ b.b09ĺZJ> qB."Sˋ9†DIQ@I3(I5'+/Q s 6C朴ҁ yaRvB4hT*Ҭ Prfb?.w~6r6s CO(gʬWC61L][w;JJI )8!( cB fnBZa ୰zlHfaB$wyǂ##!ڒ2#`"-JioyB1B[XV93(* %"bbȍD$249bH"if9FΖrvrvd` SYkOxJ5&YP>#v(ZtvQVrK#?zJs: (M &̀j\;:.Ikje3ܸwJ6kfsp<8J,0DuP2㺸7z/ܛ{w)X0ta͜XLjAZ vN1@l`UzZ?ʗ7E]S'#7ގ'֟t{MuL[E~^M|u:͋rI ,Rg2DO{LQ"! %K9 Q Ƒr|+ Ϭ8 akSw{+E Z& jnQ`GW|Ow|Q+#Jl gJ zwdj♅b0#^WL+Xz[C$U'fsn&mK2]7Z^c4p7aܢ)xtXi U,8BRh>循64_30#O&V-5zsЬ?|}nZĖ̺IJd۟X1@u=  ocUlFeI=ÚMq²'N^S9GnCczS5jP^5jw'N2K n|ur,6'Mw) kMݟOZ'T2qCZHK7PVd_}R$:]vSsixU+<σ?19VvWYg k^G^%R9\0$YIЌTSBHX'9raHjXPD*r,ȍ$@R ˽);'9E4\tbpqrɹhg_e̷:7?'*o Lc8ãCB(EEp& EFd>ZRaS\-&x#32x5sf"andFfd\o qƮX(2ch,Nowi%pEPxus9jP^o_8bBia‘@R@5S{LӞ:m#dr1ˊT`grRAK~T1슈*3"[D\Ǵz>peтs&8lRD<2AK0B\_0 [g($^2ΤfłKsI&9bˌ"S޵$O30CwHv. -pAĈ,y$+gU˖٢lɡ$$%U_}]]8%:\gkZr_\-p[9[eT l4 2s$jm4Z>p`xb[qO<\Ni3Z?e#{-+Kl3nI@c~:d,!|89Mplu_eB/17$`V_S#mUO)a r CqG㫝]#Oo0u:N_&vQb7)bL!$Sdxƒc[B;8s>gz5$T6zW 'W i`SEpo χl [WaCvCէOi~sM510E|r:Nj .ӱ'*2^^ې/yX\㹨Xpy\4t/Sfi=IS؇6",ӸF/JTJ¸.uՀK>rc>J|qf-ԦLX7{`Yn5 .%eK HcKUYd% 5ff#4ܭ\A'R[dLJtk.,~EPE,J>/"f3~A >i,8^;ޔOZٖ!J,F]iT%ێ~vJ?єp#vpP€Z" j0jBʒ( PjV'&"eNe傩LU bj"Q2-ٲHػyꉐO'ύ7zowպt!nzf0CO_LJ^LzmN+"6K-ZlpqpPݚwk-x*J1;vI -IJ"B8K6sRd*ɭ5uܱSF{=3g}gוeR~AC r-<|G)b}e%ZFO]LJι/E(%#ֺBwIѻUE9^յD޶DouүMuH_vz$9EJire\il`Tx*!rϬM;k\9v]a g/F.!ppIx!1`B5\r3/ $QfT"lHa|05u)ȅ h<2) jʔdĶ &Ζ s=?aO#%9 %j[t6-ޮ[mOmWo0pᵬ|6)6f$؜eZڈ(Z3q3dE @ʑoPÁ2 \5Eq b`Z%A6L6T82`K1"CHK-Է3BcPX% S ƣఌE :tcsJ,lonmB-ܦhI;X*.vV8ڌRszpUo>>jpf6^U@ۇ )OҾ+Xe1lQϪGZ @sŊ5^5FFft_,l 4njy?:cFa=+w餱/bVA_T:9 ~M0AqV{T09+Q泷9@@ x@!ih0 zlX#tYCAy}!^ܻ4Hcu= !w,J[H2ZYtnI'犵! m,U|Ňf|crֵƼ5z|,ˬQc͊ȵbKWE[ "%̤˽Z+=y*橖*=#Xf!G' K ef49 i'…u2(5p@( 89u)D*OH'PIX ^89hXk|!P0I+=6V^w&WdSCJke`L)Nqf!{kG0lHf%1nFxrى1d  c3@I `MM!9$0ۑf)L)Q#gJ̀xpb2&-|Y޲&Ζvvpv5L- e;h#X9 v՜i j:1*8nN i7@N?3x,?FFy EԹd KF H&rfjj!-.Nz\v؋e!p'5uFL~2aT+>@;h8j5x"e)2M{k-.(T%+\10ƁW7n. ,s /zr(ךmAxFq Huei!AĘ6%YҙG{1'ce]\u!Ӳuqt:A/կ\".)KBJE$0d%T}EY_oμ*n0{QE]>s9'5 T X]bayX\gz1}\ K~ =9")!GW(W JuW!\q*GW W(W(ԇW(R\}p%6\KM@\KȱJKաJ:JT:"BJ\s,p}pUIi;J1;"Bh5hPZPJ.:JsmK˙{ѹT5*Mݭ~T ^l(h' " _zVGFs/,)+!4!S AF矾sk>`pۂiW͊(JUKEL GW(>9;Hk9tB)YW%\1aO]U^ /EÛ_?ThyX.=*dU%`=%p#%K$+[ZaDQ^"|r8]:8q! eῬ S9iBLDN "QJMH4`vBA! WTV)4Itڻǀ$cT|:g8(] J0æ_^sƱRJ.ɲԞ1 2mד$r<ʳ7q+eT%Ht9XJ$z%I1nE-;h%B*޳'j .sI\ZgL1hqWB xCXKooUkύv*ݯ;Y웈eֺȍb aTg. sLsәYo %[SG{mo8+{T϶/btM1r0Wޮ{8ϭf-78RD[ '[ ZJ7fb'k_-ӔJCx9si[2vĴ#KLcA dBJ*ޕ6#ٿ"*އ~b;hN$*-ub3SeeWЪj%Ōd2_^1)ΜVr<ɐeUbʭĔOy2>Y+R 2 f S"EebZMĔ0q6=ُ&OWmQqFP{ [.-yR7m']OPaSZ$1 e!i-MeS)fZK*EӴۖbMm์5fanසmBNW)P֯'-"^K@֞"N^J«C~9$`pӣ&XQӊD=`Y{#1|LZ$l?h^>tW ▽.jې^hÍ`a{Ȅ ) G=<̙TV*WJQE%| .R Y/0 o(-]b&2dc~5q,; Z\|mS~/n#zNzl;I2>,wu.>ӭRNI0j)n.YBSD%q,XGH5gd1 Q+j5 !X) F m&3M9ȤPi$ϬRN?7͖V9xq@Փynso-xP{ [$MxcR,#0hKdD)БTs-dNd䩃<┊ ֐X}e44DJ2l %ˇ{2hA: E  q\j^9^K2rR, 9YWGYMVX RH4+ՠմγ>(e][tPQ82Oږe/ 5,!+ ^o|pa6IO})8B>Fk9-.;:_rצ\.G2Hh<^+30f);^kn2'\ >t'mkpi7k\0qj=h<-F/p2ZP~&Ori[_շj2}Xž> =]]Om}!qOȝuqE ڿtI /v,'Jk]1z氀Mޱ]v~]d˞WZtG˫myNI}{:NBŰ75l}e :;?Og37?͏Ѿj hl)(T"d}욝]jVwyL^nhޅ[FR!PY|FTRIZ Bf&XЌ&ffQMI"nH)+a%[*XA9 :k gF,gTWm^z^+_WGrd] dd좲4@ȌlR`VeNYY>{fP\d.U8S!0䂠N۔} 0XL k԰62Vg?2f)/fcTB3`ߺo7%xzu"kx<sÇ#@?hO.ҝЙEwgR7(Ri|ȑ g `fm֋9.p!q},YL桔%J(M9S-&"q)r-q)W|qzVg^}v'sPЮ !pj{:Rn_J{le هE2Z4.8kx/*Pd5(la>,(?Pק&[a8NdTyYs,y*嵒0F%kx\@02EۺU=I!JXdlYqU6J(X{G)H 8d2N@[M> v>x+ԧ|G(@#3%;^VZ1.S8*9OW漆#/*A roB, L['Z^\@18fLJ(O*!t\,SqʜO޺uƽl'ٟ{3Eb/h.-X1 2#7 Y$>Y"YOҐFnDo\xS` zy. ֵVRYٯyR o ﮽Z]'6Ѳv`YeIUVb E_1NgTwjiP~VR4͟a7Ⓘ7z糯O ꗒfOWg_~IRo(.C)"`yCpAzJf$]F#- 8&K:ݟzhC:苲 BtmMmmy ˟t.rz{Wh#F]g_ /jyum'o7ɼ3ǟuP}CNO4*W3gtVT՟gqӻ<nZ Z6 ~Y0h!: Q4JqFnw{)$i;!Vfs?cՎvOF!1(]4ȎWsIieVx vL32$3lY0LR;% Y)U dtAJ&ޒ%Ygvj读'd xw (\Ɔvװ9/Vp-v,K3ZzZ|o! vl)x|Z^ZY<4輱 QVpn'/ mz0L8ZR$9r$OO[dIǜOE҈1X:5(R]EzLA>%&hrAȎ vD Bx$KAd7BIIևUqnj\%UFɪEJ)9Qk!uϻAIf8:/nƋ5aO<}_{/2.[ kvץim#H|c;:,[:G-B4;[Jq@n%uvj@& PeVtjҙ~C}Q]Jۇ" a{Y$֝=jl~GYYR(/Mt./S1?{ȭdO 6EY OG;ɑlO-vl25n{T7Y<5j6(u=g߷jݻ;FʅB$o>>ě{>"LVI6w;0z_nE~x~0,dF2L徺2 ?]~]0l+ >=:Gf_0N7.ck8VC/}I*o(&خtFs_54?:C%d6px9۽OޚLn&t5`s2WFe%g`A=O;nZs3(m!y%%DG Cy0],HC_hgDhpE"Txyu1l pCڞT5R'>GWїIHSx>ҬV({E `dyc RRLjjmSڐ m{s9GԲi X61Hf+$% ce2Aj4P p9Ge*!OG"ao08)Js*/èXwvӹ F8)2NqHJT0];T?Q4kaL)Nq `,¸3&2!)gy@ƈD[sd5_=~vjE"u&Qp&apHޥsU$%R)ۉe OZ,׌3S\vLi’`vgI r("'~V;{fuLCCN)dVF`V8-Ht!Ya5%Is>DI%]{ ?s΢a2}V(T$LXI)'!AK**FdH6I/߆X=Rg4=sڤQ%#OR@(F EcPel,m xac(LQ3Ge*% A:PJ8չ̂3"طf P 0ֳ =H[ '?#h.r2'2.N<ͺқ>tm3cOJ/y?U*Ye0 f2U`Vϭ2Ϫ f2U`V*Ye0 f2U`VmU`V*Ye0 f@%pjj_dԹY&,?3MûN*xVϪYmucf9ǁ&p(hGz7v'7N?|<]^`GGa|^`]wjm:Q?aۇJ~:ϷZfv5|v<_OW/Ǜ67+$G:QEǙm%L㿢r5{X.ZkA y pA H!4bB"qC:&hȫL&h۬Δ&vT^)\{Gi܂s7\??w`B݉X:/*r'J ꂔ N#ARj_톂 otu"odvcCjI >|={oRf 1/4#A!AkoVµH]~L 3Jj4,Lx*x x8AE Q4W4H>#ah?wYޫzC[, KR Ж(2 AHg! &Q…c&9/r$b\țڂDQ 3,TSΞ¶sg}!74$Kv@{\۞aOk~SK3/^f %^(t村gQA`jfKJJ"Q!*0|9K⼵!10:aȓk &t"P +! "!ԉ"iSܐ@;fr.T~ d֟l4|܆Ѧ|$)b^!PNLV'f5~QBaTJJQ*Pg#Tykȃ[Wh|' mi |x,"d#'yy OOݟ?ؐVGq꽞cs6W45da1:h ZiFT/V_ݗ  ԓHQnvwNK%@<;phb?7:2D J普NQNRf"L@M^󆡴RT@)ǫ >j>%cx;O]oB0&_)^ yS^/7+B";3K҂eIAYmyͨ$NR+$SYڻ:Fo3THY"-:QY-R[)sy.ܺZL/c|ٮv}t%Ȗg^iٖ-N~M~UDw8qlPxt+wi.otߊW(:B8׳:}GFn56pecsN/_4Bv c#Ӎ$dL  ٿ̐]} BR>dw$J,Nq#ey)V`9Ƨ r)ɍ#>R5DאtȾ-LHsLdJ*a@HӜ)#E֓˼Q NR$hÈQ*d| r—_!⥩¡h}T9|:qQ-4; !)ӆ͞>J}hQt5^hpkL(m hQ[>SY"BHHBOAꐌc) cPTqK6FbٍJ1,,fb*b7^^ef!;f~נd|mX  hxtBS74%%!.]*+Á'!{60ceJ{%A3t؎IL(J>&5Fbٍn6 +!PPuں0jڕn͛CTΜDDιu.`5P<Y:_Hl$y\\P`yԈ5!Xdb~`]H2`Xwva/7f` "C #"TDUk!<{G!Jp% ,FNeHl4ڥ`'ʐXvY%@^5JIJS`)LQQQ#x4[Mg΋ˆX;{c-veҐ:Yɡh 㢩Xqq+büQ<%cL* _E0J'/$DjLt4`xb)8W3>Ja' `zXSiےjd?2dm~l. Dͧ ;JE2uhdEQ5hx$hwl O{v {^4ҴL~VbE8|G '.Rr&!4H"KѢ͡DɭwB:Ijp@'L-8 Tzs,ܺ؂B"|%@U2Bt_cڞxǘ}&8\?}VwM=@~hFeNU_^7L[acHx2ފᑉ'YtZ.qRH-F W;XMv'L"[C8w h)I4d2hwhYx_%rxr\Zύ$N^UI$%xM99!zs3>jPI~xW~F|% ߍe`xO<as7t<]Zm>EjkNVazG /ESHE'vҙ:1ׂ zYwoDk-kQ - DϥVjȝ "D锨P%1y XZw'P5O;XJvI_asYlC|8,bf[/ _cA=Ѵ]E{0b#XM[kD-"JQ ` +5cK#k \x碥R{dfL^_&&’d4WY!ho8ŒL%K-taBXO9wM5ipyКV aZBb>*+.&7v8S38Ǔ{ ~vt9N& _r~d6=~O.X{\^h>+﫟_&+"uLgdp,rѣo37Y4u'?t 47DCs4t lq*o7U̫ƫ_`QU=d4mz窭Yz1D<:p9K*ֿf*L?P>V=7+ɓ}aETF}-oheO4%g]f r`4!$eH+ŸrWo xPJpi!Sp!ă6\0(CpYB 6ӳܛ8{k8ѕ|"2:N?غdu3$[ٻ5t,=ި8Ije}`QOU&Fxk- GAW9oL5*t+ch`MrA c+nQ'}zM5x+zzIVҎ7_%u>Ev_QaH;/e;/tНBw^ AJQН`| Ɨ`|;/tНBw^ y;/t|s%Bw^\Bw^ yq y;/tНBT y;/tНBw^΋.tНb y;/tBw^ y;/tН3ُ={(tНBw^ y;/tW НRdНBw^ }8JBw^P;/tНBw^Ο"CX;/ fl[ y;/t@ &L\g~\ p.?YٳЬwrYWGFzI4^hя4BPᨌO~}zYLŎnS$v\]I{O7C2ѭ;9Fc##$>hYtsD2ِT$ڄ:^3Qs Lj85u' 'C(ZFp˨Bdhz4(ћ8{:q~D+{%:M>x|j ݲD;w}\VUsiiB^Zrj8OzǘUKZJ.S̑ b 4ܠsIkc~8L+uiW{@ iF^' TXs]yQ#EEˬZW@Bc38 Zjd<c+>ßl2U¾k2ϕQ,N.`t^MRm2#_FCMM{)Ds)IG%)z*H +@"KEUɄ@6.,O%@3Ϙ#cYBboŤܿ58]̿G1! !UC8)ZЎ'm|8J%?x!N~t 4^{}5ao[%}+Gw6Ξ7s+ЃXJF?؋7} >0Ȉ\Prps/d)_tYmU[oM'2Q/ΜMs4r e4_u|WbGI(԰I o]oᔿ&7A6o@YRC(97KgimUmO UϭfxU}?;[՚P.: >\>^nED+2<_dmE8QֽuMt:w_ [dZhLaebYO<^jF?f?5]vuDu7]s+wQ_$\~Q&Y Oȯ/QɶoyZ$78ƭ}uVDbJUy5/.mcMZ;=ƣ귳g՜ظ8l핟ۉ1͇Π}]}Z3wYu[MDE=[_w#~iO7tۻ`loQZJe xYCh @?]u磶1wŸ7,ަꦿtioŏm>ΓpRT@A%; $PP Fr#'<"#9p9- RzP"rwNA"4H"U/T8_MXOcY4aw]h;d3>CGĬ-c=N> J2+߂Z&76!#cǓ{ ~vȦ '6*_r~d6E?' w__ _e"qDK﫟_&-+"0)JJP/T 6oAףUm7Y&I99kN~r|@tdm,!ۦݥyEm#<׶"Vd96v:^v>;߱xg3OF 98vA F-7NU8KƖ^L.OnGsBS(!B'tQjfq5kc"ģn݆Q&xo͍dmpaL&H.'0֤%iFJnM<!Fa(N 'Ҝ aԛ87{نQ섓^8 RYb-x@If<<)*VɃ 0TG'Qʸe,wi ;c"rTlX\4FT3gV/O'i"*(\I\FG.]X |JL$Yg*g)҄% gI r("'7qԳ[d- T<},r)̕N ]HaYMzҒ)'iR=ۤcQ>USN?gY1t~@ Er%(.G@ IHRzZ~6<5fSoǽ=Rgk +zB1qhrܕ?IĢF*Σp~oc≬xic(d2Jg@UiPK81:PJ8NԕXf-g;zz,96NjCx#Vc q#dMi!s^_}^0xO=`cюF]ziX5%usI]  %7r㔑 VjZ=tRRpBm2듓D0)%<%E[J II,EJJrLCNoWJC"B( #N6Y+OY WI2 J2љKX}$.;cs:³hEofzt19HyEclDLHk Q8"AHM}{m݂;dJ8@)o)a`ߎ~z$njH{g`X@&sjat̩f<B>-H宩dO+ݺ4ց.o)ޥ9RG`"Rg-C 4q\Z> F:Cd]*n[,uc]rmetڹxf?Cc]Z.^O8#,>ͻCWԿ۾hic6˾v%]M5۞ŕwJd>NeZZom[[&ږwxd[#!sސ֒wZk_Kv@[(DBP Ad"A uZ#3eS21RLA1OA'Br/$FX# qu`#ss Az"+$7ON{j"">< m7EL?ORoFdAyX#j D :ӭG돓/QLRp5ٻ߶$7V`7L f,i% UDےm6''&_wUUxpbE `e9GsQp 9kALP/%QDkeJ|*(1RMebBjh)T+B4a[a6՚K9S10P)B9˜γ'&ͺx`h|}FXPb +r.ۜgٗ˺|uz)!Ap4Id\pN5X T$Dw[zoO$)k [wW1ޑC8W+Qo)öel4!QQq$fdHOEO#43vl#:(ymDߓm3^tunڠЖ3g˨*029mX!N=3gHSQ…3o΍ |LSxb[ `** Sqښvu8C27JMraqX}I/a/ܣƤ3p p23+K$8x|a2qgQ A`ifMBm)$Q^uyv|!$IJ\1>a BPL&"I 8i\Dn q&Mb o3L!!a$R|t:橅y]w~z.AGfz1N$lHm1:LkRE^0mڊe1>W]-}vW@ BBl=@/A؁D#Ox42#?9k6 ΂&Ĥ֘Sn6VJ)I&tJh.6G6rI3Wn߲OzFҝ_,-rK&>{Ƅտa栌 4ՅD * f|uC<.`:^]mWP&gH~fD$"zcʴs>afҺ $w.QMk ״Oo^&(4B@ l5%&LrL8) *j8`'.3i֑^-q.M7=$ۓU_xjI+nu-P]7ti?`Z~r6bnLd-dBzWN$LߟgCxS8.lŽ~;nP;^1y\7EbsJI{ELȰl7Ptr}j۫?M`$ ݦZ&i{x)y,/[,M0P97l z[<ʿ~;Q|nty8^m4_7x^ ^}#352^?~|5Z?@l^E;޷E)+cfl2F%҂'#]Lf'skv=𱒚X&^Ll1xC X~uނ X; ~>̛b8Ct18o9 ^؋ҦL)80%Z Kbp~+Ǔڛ;'vq0j6㧻QO~dI Nj)bs2X&q' I_NEMGYR6o([ ?iJo.dޔM nVQate[Wuh-ZRRnôն1XScf=rnW#MjM+ mTպ.L&ػ_Q}+ Mm9wPkŅ\qjoEaf8JCˤ.ѡYM%piP5)%1"~R?2(dQAa# T+k2({EY;\إx<r-֓JPg,8Lx\݋^*B^8pB%K͟EVfZj˴,ӲnZNJ͉4V\\.7[ݠ +ZY#|2=⠠O4vWg$v8"]#C%4V z#(E H,ﮑ-9z|~R 04?D1D1VҦG1R.yQ҂5]UW/8RvmLqy] z]Y/j8(9|@*/dQ$M5G SF fNQ@HU?;ޚJ)>ÞW8M&_WbP |h2ZfT^o2*z}2KҥPa/myͿ ShIޚ(9vsO2Jv R;M#d\R !@+)Jp# p4?CR&CJ%C4=C&#9#ѕ<6 rB爲utm lZD-t=uEgb&oψN46_Nt|EnT\ _t[2=\T4ִ5Qthi:]!Jn:ztŕ"5tpn ]!ZΛNW2+m+l3vpmk+@ki Qr3+jU0n ]\MZ]!Z&NW %])-8M:BV^,;e֨[4nD}/9M?CFfZDWذµ+@KXPr3+cREt5iϘ¥-th#J.;zteլMcnϘ5-th i|0h-͈>;^˛l?l/ǥвG|ʦM1=JttuB![DWبµ-th)M+DYaCKW؊•-thM+Di;zt%ӴMtn ]!\mBWtBTvt J),i]qBmKtB#])b$i]!`I[CWWƻB=4p<]%ڟ%]iQyyD􀳁'B["! [W^oIVشgFRҚ1;D#J_hдz\٫2+7.*VM B~^*+i˽a{i0ƶJ,5۩v(aN+汏@-k]D#NY3Id8Og^态JUf}x[@k\u]>U.ZF`),FN8B3l >9zGT_@> 9__TmjVEnEЈR:D3 eO TgAbRkL)YJ)\, Tcޜe.$y8LWPQR߰Weݿ)%Un/{ yl/{߻G1Z^RFuNgB]jր}r5N U-N m=4sJSHBXՉ0$:8U$jΜ?f*<=O8ٗeȝJǟ. Ǐ Ʌ{?9/C{$lS.6P=a\[L%,r^?tgмRB^n*aEF󢔛Cn4wH%3r]­{QYoy7Vػ$ ؽ*F'W|~eLJyXۻ'Y\bX6)Jъ7BU*~*'?ƓA<єp#v$!LdJf 8ʒT3j*+3U:w8X7q.?/ AO tT)L qj4rTnbN ԤDyb50>2 K>d8vY| R̃(Eh^\0!LWׅ柞eq~+@_LFg',agͫ|rqyecHV_[BW;iEݞ7.y>WW+O|p/{tC*K/g_ <?͇|vH4Mbb3r>zu}d{srU\A -YbgU*ks ™`wfIIf<3Yp.INyjo@?M&lYZۿ y:/޼[^qB9LP*HUmoʼkob-$<Ѿ:jZ1HuqC@;v"&1 * F# d%+yG qqyye zN2}BX-8%s O50f$vP}8Ԭ Va^"ʃM3~2hb"11xIx yn]HI&e3mmN^qٻ6$WJ;RwnAArw1~ SBR!)RH4hyر浦駪`P=@ⅲڕ,:&\s7ẎBĵa\njFy6qL5S=!?s Ҡ`=y&gzo!iwfCl1l=@]Ϝ=*-槮|\R u8mUw6k" n9+R@Iw^ṬQV;(uL^[c$Rq%l:gU chYjȜZX)ñ_2>¥]&,xjzG\K xj8M(?x0@ƄTf1 wS#G^{u}734|w*k #lո-԰d nxu5rٓözC'WޛT}Lڸg?D/G%"f9_m$tw@[ZQ;xݢjZSgzsPS4Bx0Q#Ug?u㲎ނƊ*'V5XAy\#Hj$L.kC@87af"k3Y `p:N~B,fUIY|К5'Y`Ժa7GO1b.[4qefW狊 n*ꮯ>w} D{bGB2y*}18 $ST1INn  nb-jh4=@2L뙘uLbZK,LeתYd5 \R\\3%a$^^<\׮\p&CB2=޾h2A`2,&ˬp< >40Vu& .atCHfB{DBwpJD!H9raB$kr,㑝㚭P0r F. cECk*ց,N+ _iy|9uZZe#wYeR^9aKeIAoT,R`D` `N^(u⶞ݢ>Q̗7?~DI8P- ?K.l ^s=(_SMf[ˎ%g}8[^YzG_}Qz`]b+J np@Vew0$k8=9/zk6Q*b8'H^F/[?()E|7glѓ: LӐڣ[$MaZhD>'bÐ"im2.ʖBGq>~@BXyUfAFY'o/yׄhk~q2}vuӻp6;w_ TB8&pu]A`D^R} _^k|7|~}cº唖:] Es1vzf_4Zx.o*ovٺ fyNNdr|t52 \L]kg m}^ƶApzJKVLCoUge j[TE+-N.ɻjL\u]W]2o?L}|>t>x1:# 53|k?8Wlw;hA{% J,Zwm77Р~_h]OS/@T]<LV bu[M̿ 5?%csJ\ >$mpD;㞋ˑ|FKq2hw:z27]!FP o498c:)8/C6{6- iy$9` 3~IJ tsd=Yxvm:| ykrֵ.yY[xUa>1K1cJkrJ*br^19a&w}V-84tq/HpaI1,J'АrT쒶i"/tKlUcLƗ2]@ iQ9 VAyCV-v]e3qsCϞƗaUk^#&ixʠxCZdP%.a28B)1LdhvfiY,cRTجRH.OIX.UȀQ UvY )LI  X&N41A1%i (wc;Lv5eEӇ9$tCxbR`5gXicLd=qC#m؛@&pR;3Q&)0 L1% +GAj2h#K Q[SΨVi~Þ,Qzm (⊑E% uҨ 0aDd+i8gb|ȡ.^g>'DZg,B4,8I5g!DP NFF6%;֛Gw"cU]ut!:[ 8Hx]R՗TK^BJM$J2!O6-D/لbGT(%%H^})T>;1͔`RRSR̡hq8))HIA(JI\; P&X /{-bȲK^qV-U:OкL`L3#5g`@.H=&+q\h/ȆW$@ (nnQjh@gH#Iio3(Tњ>Q7)Ho)c)'BRJ'OؾGIK~]~b-sZ2SRo@^Y= n:Z꤉$"Ft ?!b@_c#fK7y6ShtF^5_N4 ذ@:n ,R梁|gVt|:5{B/)"0M1A)6kq<,*iF㕨gt`_z)g{$j4湋רTb  FiE=0|;>d #B9)g*L׀J#QF8f]H$6)B%W J^̈ǏmwO;ǵHDg &4 Jl( A ̕}("e%IxTSWdԠ xt1 ==4!RFR>LzzWEr*ri]Ca$WƏ>z~#&҂6f4Ey5O _$0BY]ŷV_яӨI^w^S]$Yٻƍ,Wvݞ`fAAPO[۲g0}o,٢$%[j؆X,V:> &ºm;'Sn&wfqBЫw_{σ7r);Jw?*J-^ ^7o[wiX =췺A>+Vm7Kfj6{ûح7BۆVɍ)7<%TGE- maZD7GaÍm{ZċN޽%7)3X%iBH.6/NReW,g9J[i)%<6 +ƒP߻x^( ^Xd&eH:'AE9)B#Ʀe2(eZ30鵌FMFSyrW%"3ͦΎr 6ft2bgힳ+CLc4䲚;ܲ[[ 62em-Uxwѓ鬖Y.nF]15m+I^ bxe{uѴxҳCgo>;cڷb7u3ɏsv%/<d'CaOq(ڛnM&:qSo(ʲǟ7ݵK9Q^UWww:;l͖zB_zls4G͍oLF3EUL%7LvI+ "/fr@{c]m|DQi'Ts)H0 RF|s H5ɞd#u $tSaaSc]&:'j]͋hi(; rqr'ٷ m|rqDN^`΂Ȣr,ycxtHHH4F"0)N"P2zEX2EzSNq飶ěU`̙0 s#c6u#c> }PdBc\Č;\;n@Wk AaFhX;r6J%8f GjP4M{0JKR M*B !{D%$ؤB:`^e) Ӂi]IdLcAlc_ԖQ[3؍~3މ@)5FM 7eR}53Rc|!5"n8m hG9AQ"{ ҘxMxXǂǾ2#Gč>;5+3)` ހj%t-mpp#4 y%(QHLd$&I>)wXҠK0IXfD̦j X=i3(uf\=. `i!N ֒GA{%/47#"T *`>s:[\<kt쉇9b<<et>cq\>ޏޏRհNtQg_I9ߑ )}$SEQ$MV:Ց(ʣ QTGѨ@#ΩcRqĴآI[gYdB#"/hU:7‭i2ݖ7iT8s*:-5{4 E@4khXa.yV/+,Xd!Sg߶>) |?XE,`"Upm mt[pq充ƆrV^y͡e~_4G1m9 >/&mi ">Ii|q PuZNAٿ~p (L8Jr|*p*iQW(ZD!"OdUҖ=JZ +d[D*)URWS:zv\z8p9">u)FWICdTpJ+uBpd*Nf*i{p%8E߻p%9+qMWOۡ\'祓ya՞t◜^":wr2kz;w0(J.=ϐK{?6KnןoO'*~§F&沝]q zAg"g<6 ן 챛^3\$DžU!fH FiOA:jA10!YƋfTBJNS|7gZOg{tS,1>X |SzR^<2'Ԯj#FJcF1 )`)$ŤQyMR"߼^꫟,T1!j9b6*AdDx尌J"< UR1pLD[%[P#QOk~C,༈gj7Rìh`erob'ayҹ8Q'u!K%KXiqW@BPH9 S'G=9z_r H$Î{bZ]r(2_K7fZv6Χ-sLꡚ(d۴)n'4r1PКsn4jͻb5Zwth6\ms'i9kzmx`xXrCE- maZD2WGaÍq{ZċNޭ%9)c w7d9@ؼ&~8'iE=Ei𒄉^+BvҴJX*V0mH2.JPB"[lďLd!i0ER2߭9,ԧ u3iVqV+]}9Z3>ރjٗ.Uܽ__ !̓׽s:|~>7_훇׿-} hS $<\B]Ƞ֣M5_r.1 F2)Ū$>UFPAOIQJk֌9afg ]c]])[7ؕY\_Vibe 퓿xi|9~! xab1Y, 0T0-dE{+D>Xi=AP=JVUmj3љUȾo' eJ"95vNy.Zw}c= %]1!5XRb oc+FA=wQkۥXF)Qa؋NX|Pܮoz|wstx+y/?l؆x.?0pF8^~kooL`n?f-M^2{=oTPɶ;^Og.;Ь3}_"y)U' ]ф_͍oo)FCpRDTE$sTsENf-u:')rdwmOB[~HYk J$eTTL^TyK9B/ /e+}eoү%6Cz7*V<ۍ} h kU[C6VY*"ebz6Sf]޵[.dj)I"te=^FN~>䌷(i2Zu6$C6XK!!%_^@"H5;*IyPㄟy5z۸~. (i[O>9P1Ҷ/'3Y }-/s4!YtI ^A4P,3BAYY#!KBǩd"TJT 5OB*2XCw%0~=v֡5:l~MTfcc1k)1gLO2U\>#!rVХlkuB$*b *C+[TIHQ Q\P.SuvhHFyc%LXtATZA,\XDh ?sd,bb<兀Oܯf,xvJf>d\_ܳ-퇏Vp@2"b ţwbW1IWτ"ju"j6Ț{Ҏ%Mm m1ڥ۱[i0lBP"X )(zMN# UIuf8q̓.;<Ȼk߭| >"'w_1xY:xv#8gnw(r0q<G =ݹڝ^iu+P^m[pן%H9bgCI $#M@zʠRgWSC]߲}yl KH2&){UE!a*ZV1-#CŦ޾p1i zn5vYgtdh"kFZ[r =Ӛȭ7#x-5P)L~װVIX^>`G9D\/WA<}jDzl7l|o<&rhY6䚥A2ֺLQ)t&YQLZmBt* dC~]{ uU.')$d<(o/AS joY"Yi[%w8a,{< UK҄h)g[Xatyk(*i+)ql v_uZ,/4p2MG $ efydbO2Xeku4( QJYh"Q`ԾST$/004*)t /!*ҀtBsnj\221?cC#G4S;w?$3Loף-WLgi/a&"2O@ktO7<" n)ZD> \ل4*ahc(59 䖲Z^U..3[goM&vI=?oX}nY?XѮ/eK(|\.Z귣/AkGOH_狕X̻?My</(-~yW/.6,?꼐Xo|Mb!s?kX]=6^߸0i,t/]74{}B.kL~bF+'roo}S}q\k7~2˕ʇ̩]2u{EFV7vR |9X<OWӇǰᝠ[38Н؝` u5<[}|k[aYyM>z,-/Oo/;];ػ6wɀG q3iߎ>)<4Y# ԳkbxBW7B٬0k{Zb9odm?¯Y/ytf]1{qiv/4<e %5%gcH.F & QXb(}Ku X `pur]/>cU z%d!6G3C&YP5 >\bqTiA&KQ`FB)cR.,X6܎Qͬv!|q<+49 Ŀ@3@(3$:DQOjR ђi*g( jr5+I0'0yAh*RHZ`Ѝ9GENU`ssCVVFPh7b.xY"E>\j֓t|9-h)_*#3Ig(])g4NJ*VIeMH& djZ͠VӦMoYFmlR2CyOu!&ڇ+h8jOͲ`ccA2$-,ռ>6y٪Bqe60m5͂&sf,rnwI&րB$HN/$If.+#Hvѽ`3;;=yɄiRCJ>od%DT Ū8C1d-9mL,%_`HK^ ѝxlfRx39T:c\Ti8_Ҵl&HԴuBlL&lT{?We '{QJJAhF  H%Pϔ=OI}KI1?))*RRJjM`Vփ[NA#T$Nqj' *qu'2a1(). D^ ]xPNnz9ǩ\dɆ{(6I4[S: QN &9iJڐ+PԕklZF݆ GR\ ˜,:HT*!GkR"aG-<z. 1 25 X( JF)9ZU@TD$phGRy\ Mp)cc :$,1P$(3C8!NVlH9>3ګ]pJhkSr;4'@u\oJٔx'm洳||M^v TjhRm\)*)J0oGrw*=NF|^ߍ)J=x~D@!> (*d,OV`R)X&o%2[rNF7qۛjD]NRCճ<~ˇ<]ud~ar{hf4m^jtcq:2s [in{XOuU-~Wd+@-ܷ=pk]klLH/DUTDcUC*NC^(i-ʣiWUwuULר\H]1(ѨJ.cQWZ-=ףӠJ0ѨܣqZ)TU䠮^2Y'H]1Nju,RkmUՠ^8}:lPƾ]?Zr r:ښj)YaglfM AAFG`gq62D-#уue-Ǘ獞L͇YQ  6F58AO7;,W@z "07R(#$sLZ$H V" N+Օ)e(*O9bLpy_4آ #b,2&/HgPEv\9߯zTҧ,Jwx`װcӣaǦβ QIiXB&0.B94YY]胐ޔ!I5tsMC)goUpES~thYt^iߋB6 3ALk<`lbqyꍿ'ʟfF?vFdCd*ɄFG c(IB gogB bEPAdheJ*9R;!(֋dP@+ S) *YF-6EC3JK[@z+IL2`6 $*w ?;#e"o/|>9?6|ZZkSC׬lޏF?}5B& GRL]6ޢE#)6*p]n.G m=J&DB%d¾sRhP(W9ߗ8|6(=o>;Z;=>EQW 3y{DXm~h|\rr MHe5&?i|87/k{8ڃ;nrW[Ր|6fS?`XQǫ)ۜ똥 '=I MTlΧ̐:a4Xw6|>L{Ofg'{钍 Įacb|Ǟ^"()hD1J`e =zҕB9QeSf阤Cttu&]k=:h(C&l h/{<.'Ye!0hއ!Ui3z, dTIKl\Py,40Ho39d@TX\Xؓ (Ч*VV1*fY<AV.>%Q2g[ } #C'^42 e7\KLJLp!¤i"d5ݎhFk ˾<Щ{w_˟~謆[R.t95@{Ũ,/y_e6[6n=yچ1([H꘲_Nj_yEAl(ZF?HuyW6V0W41,9 dCY\oi+o, NOWo]-D5[8sz:dzU6~U+ݼK}߮=Zdv{7~~$U)Fnfon4^,.o4kL2[ӕȫru7-`}n][X휐o]?9Vo,.~{MFf9[}JMzo{LcӔyRFG1ebmK=U|Aoq^ӗl״`JJmDE䵷d(>`:  BD=Oi>FP~@H%B4] IAx 5phd{.,ds bBᐪCE<$-~h38l':}t?:xghNfhx隘W Xϓ#jP {>HESHb]a0YUJ{2'XhmOZႋ8!d/d1Y{SuQg|:G%: D}IVǣITQ5cTdq*uHz$Lp( g  v`$3mq$x)3񙙊EiE̐T(DU:Q(JY[[&ND!夒f :!^"煲1b؅Rsڐ :ȹP1[6`T/r$$!VK(r=Z"T.Tj&X8-."Х|~)R2D1d)vR1&6 `̢&T0 jŋIΘ'=/~~@3VVWlwk2O5>K V"oeP(^w/l?F1{0K&64H."j6xت#Pl! #"L[ o/Mű{:Q̠Փ+< !:DT i_bɋA< k[_39T:c\Tir&\_Ҵ4K^B\ $j:!aȍ{8Ʊ fɏWUk'{QJJAhF  H%Pϔ=OI}KI1?))*RRJjzM`Vփ[NA#T$Nqj' *qu'׮ cQRf]A 2$t+rSץG8~iO lh$S*7`u.@L-sӔ!Wd+ Cu AI鏤H@V1;YteU..))C0(UB֤Dj?V[x|cf'acdjPA{qS}s̉I(xA鏤`zvSXA11^+/uLIdYbHQfpBB&!둌s}oC7gW=^;֦Bvw0h>S8P1Rכt6%f0w9,_4_]DU޵6qcٿҧd+mpkk'Τfv65)K֊"Lߋ&EJl"AۮX D9}m8`UNǜ*FPxχf@R w'_;'4=}z바w KlLD1z"\0H(]"ڠ m{~K[ Kd% c&Aj̐(Bsac&97^0Iĩq!Gl S23&$0;.֜mW'|zQ:%1tl^nIeNx>ߖ޸+~-< S~ϧ\%^(t=IgQ"0 YKJJG B˧"$\ rÄ %OTy|$E HHDTm"i#)!>!teJ?M;Do_Vc\M"Z7xW < AR,*ʉ)t<R`E [F~PP)J#LP=d1 ==|iȃރЊ+|' u:b>Ԣ2# TSId)G"ȓ:a<8eG6)ŨXO "-1 wJsg5$$L{G|_Y3 ߄nŠSԄ` r뵱"RJL2f}fsfg㛛l3il<'f!oi2\zbaN]]/Gi%~-(<6uGURvam|X8Y"L靈xڟȚ2r6=ro)RGnB+^7wժڱh2{\~jz OVQ-Hh'rFz+zytQTbcc\D%r\1C3Y*D,(OBr`9"U1UlXsv`_x6mGꬨ7w ٚu\ߡ-6MVmd5Yi]gV2v޽ͬZۥ[4rcܺn~Z OYn6˛[{USnglk.W{|_楒dYɩ̑"2:#F.& b=o: R@iX5g;2*Ű Ua,T==b?vvqJsq Էw~Spyx9} ƂXI$LHlRv9&m٨$؃wqa+pJl 161a*צ0bkvvQXFǾ Qg;R9sp.I ornq0~!IDn䝑4UQ#քHbS!FB1Qbَg`<D,6>EDSM="n R޽#qY% [ *i"LD] 8%채5 x(6)MyЂ3F9gj49/ #bcWs˥S,6JE[m=.ndT)YeT m0 J.RqtBHz\lqdxbѱ'.[|ö' y87֏/dȘ%|a8_kj7#/֮gB`xM`D"õH"IVVQf&qsowrIkyXrJ+/".Vθ#ۥp5i]l$WLr6/~A]`\re2bM441sIH. $!vd"y. ZSѝ.4vQ8DF897V#IsAc[%o諻'?+-.MׅBņp7nǩx㵬.†A✳LKs1jɩ<^Rrabd*Sa˜CJ7_A凷]Pf.W3"ؤ6L 6TH (8#"D<PۓD:e[>enNbռdeӐ!SEcVnsT3Jbͦx=Jme~·ϵ44LwN#5ӀIG%)-TcCVA샫GVb&dDǼ.&ںRK 6H 4Y8Kc%* UB똔oB3d6yq^7BCuCkmH_e࿒^ x&O0E*$%Y/3×4CRdSY]]]"n:U>3I?~`LBq->Ds=s P(bxr&4k` * nFӹεřG_ɻtGqĂ:䥹LIL`сws;M dOy] Ÿ:MJ/J0e\2yv4>YWа9>Z{Ptxŗ:L=K9\&/,F z7KS̍?4uR4ñ0gw;ވcxčrRWB.XȇoP c=ɾW1b\o7kr!2T?FB;uJl7@+lbbMnQpȭnInG-g]u^\l+=tj3J]1?O(DQ.ƫޢBs8}qħg~|;j* *zzZ]Y{zJ@۵Ew֢mYZݷ˵6:~ɕ]WEH#-u8Em^qWǑn/mO3K1`R<*W&%V;5F3-46hut`# 0^W)}O=j ]VSxJÎ+-qo& 89S:2Ca-) !&#%1^hX6 qAgb^ON8;"ӝs ēQ`;g"QjN `D6aD.Vg^1jSN )@d[(Ci=4F~y[ 0vREIȾ#s}+4m ScCjR3d Hy@BlL&Bqgy lu|)$%*⩶dR(MԪĸh9bg I94{75$ Iً"$%Zd>ŋ FZ AF $UY.6N\4p4oga$V*x(u& Py`# E Lx7fcgzM,x8pw'dC~7P'Nlڤu Qknr)лq %zAr:+b"vT D=i VRBl$>Pmp{o`Z&nZ{Q#2ƷS{[Nb6t&I x`r@E$s[ |nluͶGZυ&~Yi0?z:-IXUҍ'>@]6/|>E >L]Gxrw+'b1O?*IAin5Zsnp0.J)05DJZ۔}YjXcg;_8^՟pT>H ?,xֶB똬my1,lnQF-v7O~;OiQ>̮'i|ϡm:YۻWE h{aFdHg n|=Yz5߀^Ck>a8Vƴ+9ơ4<0$Td eJ9SUPB:YrzvSMQ7yse459J ^lVCn!vV\s99#:ERv4t6)g7?~QQЯBQ0R`D$+h{pԃGωN2O9J )5YLg)Θjc {ycd =!"}:ebhCI9 Q Ƒ(; / F7 ]L8U#ژr^(ژoSc)υiIFi: 0^PFqKI)C{3 _Yo^AӼMe4tXm%,WOeଡ5+&FϾ!Ag\90=#XdPI%BOYi ѽ=s[{Y Oxq/04_?ː +4M[J$"L[$s7X@Qb#FITVtL613)V֧lmFDD b cpH:3:Ύk\ &jS:%,1t x5Ai0ݎW{kӧ9^8"=LB)u3eS[c|F `F$ZYdM8FVDkh$˧LW _bFf= 2Deۀ2\@Hޑő6ޞ';޸\f H}!"kVhLMOb(ILN6)g3hTA .&RM:YPgW.WP,ХA& R-aB5fs5Wܴ-pO4 3tP|>xk/o~ʃ}L-udPzorН"a0 )o#KG|ъ< .Vm<"P#!uC(՟Q!w"ɾ$C]5I*o 9hjd3{|2gI"D&)XɇamgZv`aM";Hr$WMJA|}Cq>--$^)Bs~ӴFYnCrݑD;uߣNw.vdL?jG %G(NrIw"2G]r֛9]A `ef>_VZ9Uީ5DEbmoس&ц[''wYU{6h0f #W^l1)#6Sfms nHm%7JnԴ.ދ P[Am(&4DM+g:4w:h+nb"˺w_Y_<ۼ,Jq+U+EPpl(2dRQ c%AGM`GNM`zлֳNO,TPLbs`vꃊi%F*(gSIdl.bĢX123u!S1{E% t񢜆%{^Ś>!12“U'\+g1)1J)Ô} gJA)*4x*msM9 F$QK-9+-axteeɩe)x{Ɨ['Ëep28[ܓ̓4avսJ{HLJPR]^YWi2(LXm#J+:G B,}y);(s?@cL"7ʐ(Kb1DH"R`*p6P`R`$ )[X1c@ȈiDk45Ddvfcg,:n~Yg !vMUn{dHge ԁթ֊%mv fޝjVlZtk`=&7V߼JܕW&O]Los<|".gva޵uЧ[ .E̾lbsHEW-o3yΞofgg6~yz5ȖO^~~M]ve_l Aw=K! Β[f(?RX_j< dkc]5*{<;Is$Dw pI $|nfi.j|RinQ 2+ (^HQԦ46半Vh3v \%L)`ƮFfb֮ڶֶMk7{+N.*wL!b֊7v FF%DlZϚQy+ yA&K&&H EǬ#"Vև٬{~(}ш*kDhi{k-\w;3Z+kaKKs"8Y#@ƱTwYg@5i%r虒\}%íLyBY |P5b5r_)T{ҋe>:qC]ӋM/ދEsDpFYsLr -|Bc@H06!qrzqopb-x>\#2Js/ e?>Scp1*t9 ;hWzt@\g<ΏmY[iq:\(/"ꃟw K~]x`Dh_3k|O4`#M$}A0&ܫ2ԬP40%꼋M /\rWK?gd:܍` N))J,F'޽ݍZ0KjnTo:Ogz ,Dxg7ȹJ=WjyzHbtHA^Y/^ezT9zY|,'jyv4F|65`Ѱ|eDLvhD`ՈqtIz l~>S/WS_Go0|JsR FbTpВgZ[7mRƸ 'K?pnaQܽM4n5$lvycob6\a6\ ;e-%Lbm#_=/9`Q؋ If a1*eA3|2Z,KZb;ƼI6k 5 e.r(u'#סo[L^HP ӂhdEB/S1.c0+Y=} {z816ֈFQ7P1kĴğ8ӌG+6JGSOϊ;wIN'SF2yc~ +נ.EW΃0úGVI%}<!s0ӲbL9Xy@{J>c@^NlqWSk,t&hheZ_XnCc2j)H"jk 0eT Oһb"Ιke7Z:V؊X|}-ǽ2 :&ݸ M]_WeEu |w_|[X L}'a)tL;PcA6뀶ԒѫN%#k5[ =Gd$ RbD8WK/UӘ68p s U nhP S SyޕĴY#K7G1%y~c ,J_:(ԮNgk-Bԥ6 J mQ?~f9PaP&Z," [R8Xb8u)}i ?^FcU:%fJ̅3ԩCCd&0@Uٲc g4婴,}3fWØ $;I{c`p{L6{L6u<@5"H C,Jr21좶Q85f UX42HR&NI9έ+4VrEBX5r6CIJ)/9M(lfrkY/=wyEbMsTxfVcߛûTՃkgmŘ q$J"9(W;R o@[n☧mwmӜݵ|c$'f5{{Nx3L{'ɘlI{GxTGnRh" >? YuI2ܷXfňv>K&z/WEM=4B"ꐱ0$ALUZ{KBu20.1Xzqƺ0=!,=AŰ+7Dt%X.]zLHh&ϕldW&_eX.'VGHa:eE!vQrA{t0䗔KӑqyOqLN˝R<:o?(h7ryݨ5.KFYI.z DYyȧ@ZR *_}!ErKс~L|^ѹ$tq̈́6mWe1:w]!lF膅PQ/F_/|2-_P9;M}M/yn4l8aѨXѵMbq,h|O`dg8ͺAAD<)rl::.0dL>Syٷ :-'A|t eSh"׎9J^N+A0[QQHeQqV2[9$o:i'5DrL iszhAc_ZnuKt|.W>-n5QP|rpRvN#z^2GXQi$Tᚺ!( S"LT%VI.($hѱjje5!HcnFUUi̱]}[o/+M^˥.A'B#D=)Fd~_&{=րoF`4;ԭ~8ċ.Y6azw;ވcxrW9!Mm|m"a71dd\Lyo I EUNnemt˔=n9[td[ɩ$SUjEzGW,ʜWoBsu$%ħnf~|7jQGz6Nbψ_j9J@&4VҞ L}-%}7A5J$[bHK3ߌqn['@0`8a+j'hlK2 ;zﹽ'Q-OB"xaXÎ+-(v"Mqr"E<穤܁*"m%UK* &°\U}b˗^䈭U}Zʦow!>+DhTHrPZJL)0X𷠹ߟΟ4VZIC4Db'>\(Ԕ{d1qa8X[x ^a/c CmKFZ`؂K(8P%$dW}s<@*+{¼%̝7oS9::Ĵozïh+&޾y!Fh*ܝ˾OḪ `A% ( ] .+oMRGUdP`rнp $ܥՉ>D-z\쮊CkwR|#Z7R_&!D~aǺ@rr(ȇux};F<5I?[bf6}srKdz?SQRs1" `X9Va=Pgy`|\I)o ngEs$K j͟"ɐy=ӴkFXۧHzΉ~5t/@bJnG(5  1ً:I^7QD&0zTir̓N.̬0'h`N!Yh>>5ă*95Dئ}wzSyl'D*Q8NOS|{3J͘f4Xn]L~pdb;b F5z7t6:xo:'Q6;RGFɍ):a`Z31o1HӚLԴ|Cseh?oyp9\ֽKP|;N<~,s%M"(6gZ[f2Q)T  -695Aꍂ?BZrC:=PP兂f.VKTP $ . \.ňE- cdg렵CƧq8B/KeyREƁEY9=6FFe [laGI5C)xX/B%0Tv}\٪tv%痳4<0Km*CA ‘I1kb5^!j9b6*AV2"DrXF%0!̀ރ KVIjwG,d&zafR۴LDn7Mٌ_}ٵKuvBXZ,v)ϥ"I@E%RoOȕ ^U~UOC^W,4ӷ7"E!wOGG/c ?Ixg\ߟiؒ#c;4,Xja|Qa&Ro:°[Ag'JzuQV(cAU; >HiTv"<ġU}{lZy,GUQPɆF-h¢oם<`jxs9'oiq]=n&i-|/`MjUԯ}dwgPcq?,PMA.OF1WwSD991r' {n44\0sM9l:%ϖR2|%gekOfxo=Ŧcs@}}VE֚VR.AȞP/ËTIMY3r7{Qq}7!˓,[&*>Q1qI C>+o"޺kuq4uxbȟIѴ lY/G'wg7_xk+֣_jl29x{7^ xO@R)"U?*¨YXPU|hMHPEMKXt>7g,/]4dRR9$KQ,`;'մzhF ,O OH3Yo?eHwց [:V=^10o:*6N+53w`4p[g塶جU?id` *a'[v6~.7@17:0X-RWq2^4)%azJF h_ivBRyֱ$r 1L qh3#EBH16ŏ87$X`RIHRVa eZ30dFMFSyUID树lY+_Dornv5zf]r]ȱXyuzc7f16)?٤8n\ҝkHnE+MN2IŐ ۞:OLzt6;WArεmݶ /HCnNx ) A&lMֻE/OɺEן|W9wkϧړ^/3%EWY/A/zV$-#32gΥ$8'4csPw.y+RB/'fG }{GPa1JS- n^=RN@ա-#eUOWmm?9׭0l~ӏTdl, kKMiYl27|HP}UyM]M0&zd j-pr>XbZb;ƼI6k 5dz<Juh;RGBQo008-Xk{<٣|T|`do_}zJ]zvYs75"o?wLOۃG\c~1z 4+klr . 4+T:6 4)&_^'e%k}4i/+64tvKϣ4=iמsƽn>!Nb6jOۍGO ~mE"i*e ?5RD3a[I\FkХ,p ;aHQV;HAQ DBǓ.HsZeA9ƍ΁98y'ǽ̻wۤ'8.)KU` \y/#˶9z ?@#A=Շzq:|H&%wvCR҆Gh%F5X^8n7?L>J%#)Oȵ >¹R]Ld-xiάL8_i抆~ C ?-M9~К5t-8* RՇ0YdoSѧ?#!2Zfׄi 0m&(Lgog`3 [ Ae"(%cQIƻJ]LN>K7}f%yV 8٫>>[as'=/z-Z! r7#Dޢ+à ( Dޓ4$o+"8$W]KBHNt'3|]kH$ʁ\%x+zr\Pg+_x[S>Een_1x,5gTTVTr5{.@Z BC7XOaw@Ԩ%F؀ )%% cwPHicJA{fуTIW`f!ZfY`@%#1Ĭޱ-W>/>rȒ٢&sjKvBZr'efhT,R9ogyC7x^X^f 41kDRʂ*l+TҡL$"! B2R ܝ +-C]7qRRI-?EeXC{:"H qttLSkt0pS 9O@Hdќd<( \,b/J  IdctȔ]LZzi>yA+NzIpC(>eGhܱ[&%ZV/1݌5(Jg޵á/>J?:9x7Gݏ|4P 1e\bK,_iJ[6n~Ӵ=mb9(8~/޶s/ȃXQE>ey6lW0,}0GwӢ@Vv:9I_smD9-3''KRߧe))(iPڬ$<g|7iKmfȤt/="[(葖ɼۉnza  su//jɺ/2.ʖWa>WDZdfs76+^E7F-n;IVn4/M{>6iۑf>@쿸5)q*ft"Gk_ޖ}. `麟on|b瘰n~9嫭_I\b96Cq7}iWOT7N?0&1}yö1$wGcx6\ 2b6sqoy;9v_V+|ʂKvjr螖lꩢSx{u4~;9ywR\;Kl\tp!I3iߍ>o(E}.*Em$C=[o!ڈuzZ=d/"`Wr}7A{2a?VIlu'k3mvW[:^>m܋ 95%׀9Fi &DtǑ ˁ\FK2h`:x2v i>P [Oʁ 3'yE:XOy{ΨQY_WoטN|9CNzG8 ߙGyKX5dM'ke!̊kJ)G)7k TCŴ3RzWfzTfV(`20 i+4DLL*:-a8rp9m `)D( '$) X'ڤ^+X ks7+:GΈdU<{/vf?oMD(+c.A#YR 9[`BzPax< ~j$xרbP%gvE*pxL:JqTsHʘ\'UUΐeVHaJ&O^DV &5(,1>INVj)g*fH-Jo%ΝstQ@VcFANj5gX*ZiC$=qCoR٦Aj/z=}d>(E r~@Gea\($ 52hA[1@jP+\Lth?߇e!HԶhEX10$˃⨢8zuXV'^an19lja2&uՑT ,0$3I!Y1"B42)6z"}}8r8=NP!huX`w ky}. FĠcxkk_*PY:[ ^΋a 8y7\ i/iHeBlrd)df],>+dQ:96BAK{&=OI1#zG ));PʦK\; C(31#j2 2Uqj`ETcm+UC,D.;9EnQ);=W 3VaTm9 UfƯ1&$?yLN0$+ʥNzT[)"f)O(f B^1(/<10O9RUA@TYGn;]h)^F4R[tt9m-ڂNFme_}JecRk:l[DQpsqYdU!s0226&zo+\m'4r[ӨэFOttA,BL_"Ew{t)l}[Kn \irE +Jd;)>t+ڀoݾ[[ m7nw֑6(Hk5n vqE,T&(t`C L+J.1#'H'?YV,Ӛ(uD Fo#IŸ%c{W,6.6,U֙"_P$e% YSSS=U!y7$FX# q/d4zHI y'=;"e%.7ӦvԩmEW-Ou*FzVq`GĎm=V?~wQLRp Br-gR0E#@r(Z9kA 9 pA H!4bB NZ$|BD yV3yΙلiGsֳ䨘:1q>=BfGn>jdIME[獯4YR]Ki;J ꂔ N#AR)SA84a-{{%Mۛi8߽٤=_b[^~;7߀{vj#ee3R?DNbI"IpTb M`ϜQ Nd{C C`a$ß@5x?;dn8p\<3rY 920ĘK4 UiF ج?#g3ɺM-:qdOGҞݕWR m;b?kw=b6x 4Jù>$ɜM2~0~Lf0+? WLA `x FDJ)3I&t v'k'k{Ρ6e9r%B8=v<.+R.+tn|'=]Qb|uRuf-HY3VL.ʴ-[0ݧK6)nr@ZCj,{=o񬖞}}eӿ#͏6Q })Hn4<.zד~Ԋ2kA^Qj (\$|3(O ;,ȵFFYv9!dG6?oN5~mS,9$[SӞw&q'T2^ߎ2Mfr\X=@Uh7^ѨcSF엎&mNӕW-۽6$,\>Lj%(DC)$GM4rY+z}Ϊz18)g /<{:#%G)<%R`%wg_"F_v+-WmbXx'+lVOt#גHf$ICZӤŐT(NsO ~I"ey';~Dž6sW3 ?Xv,/ڿ삶*>'n~?~)߻~mB0i&{y/Ԅ67߾{i|[.~I89t~$y}Rs' _p|/OVjOr^S\⼨5SfuAg νքͻƇyMK-"2:\N3:]3mtmVmc3ksТݺ_Fmgok0j[Eo.&hV_,oE~a,7s{4J{#lX}Znz3W5QG\Zz_o J(j[XW-Q-q]Jꝫh)hrحzQ[0^!Q td:T@Ά%/iR:x2 T7c|S}ϗ%PNS3!ɨd+x/9=.3+B"Of˒&G&qJ\!E`+֦cXkb@51'(w[ "RbN@mrpOiTᰶbjWQl9n8\hv,[ZǛ9ѩϗgzdf*.=Upٲ9Gx(W&GP֮LsBT)s'V}ri]a':GFc)v~&yݸQǕ䍖a>,l~:IF5=[=Êg;X^U> ܾ7g+m!;훞?o΅!ľ hѼZTug3ů$X te! +6G?_.qُҥHTY :Fĩ&J`9 1Br㈏DŀiE у^e3s=姉GF::*J44gfp .$A+e^JyQm1Q3 e|4Q[xkp8I4CUN5iv NbOX痪HK9-f'/hmrf߀|Y Jr4rq5^hpkL(m hQ[g}2DEa%2W^) t !9G@Z\ b=уC2$AQ-R؄Ӱ42SqbXX2BU Հ'Oqw̸dLs /NjoTv<~_-- ZQ4$NHnjXE.^g1+9Ma\4. Pma(1Z&IPP#q*  Rq$nF|VޣV0EUag~HS<^JpZ 'l>P Q*CN<EUYmN{C̏(]ѣFTG,-DTJɍbH 0.yd6o%JnSTI*VkEY,3B~oH]9J1. H_ (+uĀ4.N?nq''8FGuB>)7of?~`=I Lac |Ƭs)n(IZy.=ӓU=&Fmk~ nP$G? ͧ˦i 8C+JFEjQN&@K! 3rrH꾘YS'|zZKrH~9$PH_\zr j+ bzZWWoA67WձLjDlrؘlpo(]ۄӹ_<ɁȟSflv(gss;|sF+nl$L]o™ BJzgۙ>uܝpgpƠ$p ֭)}XٝRzZ=pތ-Iz5ƎUtfyFcPp^ ?0۫1p-&މ?<4syEȼ;D$s&!Et8XIsT9#'.RJ&*O-x 8kw k-cxMpZmWہo1y'B5WW?Vd'1 FDHNBg,hLTF ҀTlquN"'8MqWI&u.08Eb"i>; ]M%II~jU;9P%-b8ⱆ7T]>8ƕ!hXU;dR2jadFB18'Ѵ8 [S!(cbLQfIzepl, O"䝵DU *X4EcFǏzCGih8FH~ aC&u.^A`,%|T0g!]QA+#>hUېMVh}V|@L<5(BbzF HC8C\.`=e^ri9/'A+wY]koǕ+?%@ZC?: YĈ/YF=MB !eA3E&[)]UqNUjr6 b|jŇI$EU^܍S5ԺV"ωRRmu ;$ףI YO0VG[%c3(mG$S&mI!!QQ~B[+ ҋHSFGG;催a$U|i! Cᕍ^*QT3i,I:gEէ (хk§HkPeO]ܔ42f;xpN+]+(IPkCuYv ӌT"yˌ@4ZE=elEL9g'Dgb {'C:mqLH$[@yU벫ԔY@h4yֲtX\Յ)c#d5̱Nc(J[|e,؎vBނn] @(Р pd..h]icQm\g$0S(ͮJPHƁgUkU=Ƃy0+ !.}p$,JH'J F(/tcY[$V̬$޲[PRS[ y/Gi,exHX6 ئ{tВ0LĐZ*m0&܃Gv%._qvu\muLմۜFPMlJ''40z63 IlQfxjЬ(kM!JK9O%'֮ `LLy}ӽa#ҷf!ƚ0$YUfR ;( ,3AUbFHz De{`}V="Ö+Y|]w؊PD4cҀ'7$#As(_*If("ZT1"&7cbw ;]CTXd 4 أt%h#*ՐR380)`cFj4ifz@IQ{/AFs&ӖӴZ W-EB9y9ݲNKyǠB]hȟ4N!ě"A[vgVyp$k 6VP8k* HFWP F̀zreR4/$ Lčr$KD9+Th5]A=u@e(64qEOt x@$\0 6fTS.*w3H&"ˡꘅG@VRpIFi:ub鹋6$D LBFt6b ͮhx蝩ydi/C^E_^Ŏ{p_S`bvI)95 ]ќMN]/5Z;7B4hӇKj"Dk}+qm6r5Gk5?s&%)/Isz1J XpVJ/Q 1@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X *@z' (hmx1J X`4VJ/Q  ϱ@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X *>|IJ Q^(z%VJJ/P Db@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X JV b@08(b@d6/E D֚ܕ@d@ʢx@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X (n/GAwiW?~$p}QOl?fЪP#AOqgWZMW}Zl)i2ɹf_/cO'cpسtڄc:$WG>f-C\|}V]ʛggwX~ "ͯFp.ˎx_L0Q~}?[û. b7ڣv{{GΔtyC+ҤUW&αӽKkB:c ^}јɀJ2~gɷ' |T?kOk) 7p?5wi\n(fC<Vlo܇r+afuvl:Ѭ"{׵K:A[CzS?zq̍omw=-}-"# c<쩇ߡGtW_.>ߵlAӈ\/p럯OILJǟdO~M?Z8]8Z i6Qgp|avWqNsy3$ldf0 -SsB6!1}{mPx#<ey}wwקl@槲qpg0➟ KrFW$XrnIZ_r'}@غP*I[C%vODsD؜zu;(T'nk>OЎ1 n4hPp? #V|(NJK\;] F0xo"5z Tw,tT[Ϻ| 줣5Fs?Bky|0D{3<|1J̰܊Ύ>v9IfuC6$?R _' ?R`O0PVӔ'mB^&δv9 ?X>#͜爐s*$z|rqLJpS-tzoKMw7oe&!nK*P"mb]KJI "x5wn͜#Nry'_ N12='b$mʖRoX s3m2` A&< :~flM>iQJhcn$'2i7,݊.KմEޡm]T);M{8Ҋ,#f焵#}a<LO +G^1?΄E4Pz6=ҷApiz)|d`n0=d7`}SQ`M&UU2jt J6TgnBlzNHksl;#X͜gдx$夷dIPSzr!͕wJMm?T'.O(&;ccfl>0RqTw҃JM*K*ړB#.h ~ ogY{+~JcwpHvoEں%Ex[dr񴂌m*VUŪb= wx`*M:,& ܽrgh xpB$ee.m2HR^jM QȿvқoOO'zWBN, ꊁP09EP2 IBC 0U>|@-1 Fc$WLs-|'YʹL&oE*~ c[k)LrQ[G%Ps%n7fpZ﹭=v;wZ#1!Vg'Qi`"rTY-AȘAn6&n{"L](T0-s|7kW6ᐢRC+zt7.m C)z HMvPkIq /  n=-).YSPBV5)}յɺr.DHJL!h h3;hg6F0hd`QܵBc6Bgl:anj+x<0#ɷn>Z2P%bV:q&ƀ )B B:hLdiצS-hV.F pC=xBmiK@mӵ΁US)BL4PiD"N$9c>1HH'ؘeWwtN'UIE.-38:PAy)en%-D@U$bnM s|)Xxl.κo^9 ݁o乻aƸ0XKv`Y@6s ʥS^ B%&!h񹾆z1"<Kl$}"ׁ ZP B2 TCd?MCAvoѬsn [^z+M2"޼Iq[MzBMzÖӷi{c=MzMߤGՋĉiOexϖB@B 406"79TFC yBtBtҐ6ᢊ8jBhfib.䙗HB"GO&IE-`CMw{7+i_pw!anruѰ--:GYHI9˴/E-9'g)$I-%' (>~zAfWViD) "!H i!#De-k;YXiyvf=)dJ+uPrT3Jp+q6sH'7!>.kq!Ds1J'т @zA)Q+1)hY.# 988u'ϒf1=GXUN3=7IypvخQ-sk@\Q|Aߡ} (?YcDdW1bA|\_4O5/aǥ34af0{je8_|*W(ЂXaTynf8C9IYauyYr2pDnWB\^o(Qglk%ꡐf~ֻSyʨt.<;RuF̷`̹]aͫ֝?_&MS^x^_zT)///7Ӌ M(R } 6SD)2k?uԓ|nR~q2K6>>vu 0 h-+dZYbPg}(bK#r_Sj+]0kO?>Qaڮ =5]m5< ïr30y7$'kh_>'⟳zc$yMh2zCv͓+)[3L)(v!A,eѭnnђ#VSulK9ejJǣd]~\G; KOrlv O<#~r<7>m0鶹Q`loa?UR^O4('鄚SJbP@ɣ,ob6E9A۔G|m߅JGGXLPi\۠Ya2MMKczﵝ'q%Weu%:^͙h)e2$p*i025H2PwTQu[k4NʽrS>2VzVQo5)5QOؔw[KI / Q3W(!>E%)*Dn)_AdtJ4 Zٗ(,H4RPP2n j@P\0K-h˒w)KTu_MB8t+dy|]G3P؈{r7.p0?|M˿d["z tGWUXk$/H{myk="\OS:꾯ku|b7¶跽}Fh n kn%{7Q #$+[<S#47?S"I+&y;d/*5A-gΠ&*2IM[P5oD 1fP5F͐Mp0jL0usn`{&Chk eV1'X:=5@T ơj[֎[Cw@P꒪NS1qX}<2=uӕzuo"eU݉׳\$A(Qm} qF)Q05xmPRIW,!r $< <9(@}$8iUDޑ֊#mwIIN7?͑K9j$UZ);pn0?#'xW+x$XP6J.EˈQhnS)J# ^#x&5QITZqBH(c[!PH/:YJ{KV$OV]ezR 7_=sq{rFSM=l11&b RfZ[pRE\BE3m#ڊ[N ӗ̺Z4GCV!uRԶe7āy@%} !I|{7lRJ8_%FXRi9"AkʜK1< 1~J‰f;'Y>NQbr"RJL2=p"w '&lqP5OlFO]R*~nrny[GH7xVB,3*Ljaoo_":#{' <.Nn>DG@G(%MAt|i.e ]CUV+W0Ҡn1pUx}4o,GԞY !b۾wz.nh.Q<q짪As;Jv<|R}uM 3G(>܋>Ƶ|1{%LgûZH0~8^֦hXz~^Z a[C:Ҽ"3/Ad=h飭',AM;<6\ֽKTz>dyy̕"v61ڜI[%cTT\nmEe%)8޵ cBL7 x],/aՔh#D3CbI8BKVQ-hm r8z׊UѩyƉEֳX(-ꅝU.†O \ yIL[xF!/KPHyH-(D*3G  G5{%nDZ5:фxXPLLo jwz8Z4Xb#&  ZP B2 D "s9 @z'6= ̬շi3e;v,ir:hpǃ=`m\dS=y4m}7O ]vDdý~u55a>ˑ'5h*܏Zvl\W+ZsTUGZ;F7ԩGiRJE%zwR܎Ge6[6%P^"SƒX˜ʄ^YF.WY}pi٦j"(Dff3%Me`2U"ĉR*qTUT 1P#xA༷THY"([t<u;xJ{Zv΁5+.zj,f rt\?n͜/w!6e B˘ƽ1KY=SpГ预Yn]ҽaO܉Z+nVH ?^E5â]WYP޵]kyA~Fx0Lv[~?h~2N5]P |`pf2~e|kO4zhCc`ί 5rhRKl[nmF9gh}6wPnTpˆ>~j`+ߝP֮@!;"f js?IFuǡ{D=oDJ]-րc–1K l%3!D'Ȓ"ݖ(#whU1ʵjvtI+@B4=b3q2u`XE.Zg3-9/~я~q" ; w|n./xab9kӪ 3UN&8 f2|ɪ;NhW7X dl18 sO4)&||o&gog_OP_CoܲQ`2}y` J"uul$#$SG>k;f0ڱq]NY LƊ@ TP/"Je>cM1p!j˔@RD9?x dI^ٓXAQ-;=2k>K";yWKԀ1͉w ˵o t_1K^)3X-EኜV'#XMA)0FwO| ]ksc^5q.1;uYTTзwBiZE[D&6촭 B^.M!f}eƺT]8"ɣTQ;tē jo#5?тF_Q?Դ!a1Y"j؞ ;+ٶAS5dC-LfT%fU^dls46~vueص7y1,M4.yot/ٓ3n+3YɿoNqfg> ~? C޳_c>vwxݼ!ؘߤ޿ >?1geJ 1ҭ|΢9 Ɯ/9T2"v|t57yIR|Ls Zq|zt68讷ߏqk(E1H!l +Vk^Rm!Gf1EQ-rH'աW;UGI7&mUcң!)Ƽ0+U a)Jc$;4Yw5^I?m]; h7ض"UTR$̈́%HE!ho: m<4QH/Lp9/12: ㄬu!cM 37C4/.L=Z="2Efiaw V~Bل0~ܑP'ŏZ|ŤRQ9E6I*M#fSptbfFb1rWW_vL&XYXQA_3$$c@QJƹReJIҠj 4!YDS3P(BUh΅yΚݪݡfZ)EQs|Rf*2!R+(r]%c*5I:>ȣ 4߾_:=# Q'rA1A\1%2N$URA[YF &5BfP+OiƏ+t i_ȋ<ڙuԃj{}Ii;bQ'Hb[uPz QNV c#h7WY RaOwoRP%LF TK()$ȋUpUQcHSuIZ/u;yS!*Rdt(Od@TpY c̔]XuLUBK9)6n{6=|(.) [SL.[+@LkJY2Ukhv> rQ+T1)@ ShBQh]BΦD_kv; lA7  [z%5* j^fE`nfU0U9QIo$m}J)qQi(^Ȫ$C )4"hcGdckAcuG^f[:lbиXˁhuGf-Ϧ;2\#`wGf)ݑȊ]w/VsH:3NPEϻwɫ:8Nk:CuRj/*)*c`l3ekfRse7`nV~\bryzjr1r'*S/voweoƵXG2iD)0!> I(-t&t !eEffEZ8mmWyr5iji꧔'5o4|gs"Kv)1L&?/9-d*~^.g;F*pݮv[XU4X,Aw˶m+֯ 9veZlGbH9Ń!՚ӛ:/6cgmlLa#l2R T#ںڲsCnt {?YεJ$md"4:%BnI(%$!#R(M <8w/tHGަvPkfnVy.ݪy.O<])ԝhvh8GBGhA?:ԵAyh_Cp5k18ʤuN [6hG 2)FFj Bg:M"H1 m–+m$IE'{#"y_,"nYlRKI%R%UVQQU`lKv;MNΫѠaU|.|{Kۓ?>5?zdStN:c`o Ȑ-ɊJE:Y`dnaK9S `9ہ<moxwkaxGfq4jg ԪO$:E ߱=N,gNxy¬ǣ;m!n|j܆TvT_3hOm"HH&++VƘ\&)*ИLHjpMѱtbl]س@.Zj\)RyT^$L踙:OŎ)oE霠9*2F3y#误.kw+v}]>!܏>e@#j1fbP9BϴI`EVyD-ED))j'u,_10Ð)IHo Sl.$ЎeQ^X?8Қ8!s,g7WyH>~8}3vq[4]q Xc(ƨoؔH-|%՚ $E+̢G$S2:p L`z6à>X > hiю̇A8U ,OS:Qdx x̖z'2بUE)9MVY/ڿ%(afYV9Ib˜= S@Qkfw+'J6K-xINMUvfRr6@>Jǿ<}Q'0;zsT)3~Lmfi3)h/y,K|qu.L3ՎS7WGՍf7sg6)<^BYn,jޑ Rl䱏noDhm~sJ_b:/5jS"h{#,ūo\䲬Vf.w.xƊ._tG>H:)"ه޴A~ѿ'׋0EVx +G\Uz qϷHR4uVAxG$[77z.βh~A)U/~[6Ei_yf!$:m7Ԁ tdGRw@񪛔r!좿R}J}-F_ g׭OS}dr?K9#G 3؎|Իgf[ٷ<L[Z=.>/Toǃ1pbz"CvyƻiNWzzǗϪԟxh#IM^b0qJGmDžVľBwN5_h8q֓h3v/< 6f)caB_*Ιs.bP!cAME2OG1M3/~&Z@7`xλwqRH PǢ X 7e/%A}jY}7G6U]O}jT1 q۪3np Q!J5?iț~iņB=UtVG4Ν+~N XvPѹ(l(Pyeȥ>M.5Υ>JvgA6 b2erFjTt+*b@UjQP|W^q$}34m9OZVom˾Y-d=2+=UpѓBfg#ݛtGɝ-+cTkmjX*%ݝRTgw˳:iːlz$ߞ"Ws%ya:쐷:?yFin%u_[F$)7{qư{ǸW3 t攖Fb默u) (H*a*"Mʎ‰˳AdtL:KJ{ĢD"TkP) }1Kr({kʾdSP3Нy $*Y@ZhJP ;P(UEu۪JxP|X\ͭDYK/-c*9&` ֊DmǺ)7%"u-v3u[05y.V8jVV{@K]1 d5Pb͑qF!9(@_(F7PXEtE,βT^.d b-Wd!'b%3m(N"ݖN F WZbɂ*ڏ9iuv12V`5Y Q-b3u[/2vq\:'_gQr]`w"iyRXP0'g<qJ;cKtYzL(vg[#zG {n{x0El0]ApC3E?A'vEw. F1`P"ED ^`-h\&k mt!tu+\Y$1.ڵzSdPU8Cȉ!^kJ甞xOrzgޝjtS#46cHi^Y0˂@ fσWAYj)'SO.u0isFc&ћN9; @Au`ʉk( wi)X "c rIlwT$52i(Wִ BHQ{IdW@12H6CKُZWԶ;ݝ}ko([q.Ë} ^Pê7/!_JêZQ3yXryk1pSdW,Q9 ukwӕybߡd<]P:wºun oցr 7P 1&Bk&@^f|yN"NꒆR,uĽۻm_+[]|˟?VXEz8~.Ƕ^~`LQX}ƳeۃɈRw= EHHʒ)@$:c$:\GJ\%ϴ54t_ !}.BTb圍 2O4ٟ` mXPS/QuUSXC.󰾍|&R[*Poşe~鳀~mYMsǨvrTbリ8כ2?G^(qr1meww˽;] _*^jQ/?^:ok_W^~>W(?Ͽ(K@m]HG 8dT/fNn㐵Z!k$ ophufi?cIFz7p=jcȜӠ. k35KheW=Oi:[OSV @ZYC JGDZBdQgVKg6741 :EP@.Z2נLU+nY$LL+OEᓯ^I^ٮ/ZCOhRL}XX)YW⇧+֑ c1bIQ%IRj8j? t*fcI,Nr!dχbnE('qksa62u8w5J`eucLgoٹETS{<`ڏ%U 'V|ـ,(svUP >:_<{-#9+wh8G)jk%bPa>U@ʏժ䒷N+.)WH97 gi͉zr:CQׇy[{V}?:Zk*5\H-t ;4杻Vd͍U>$9lìP|h^r\&ycԛU5 gQlU{p2>a9wg-#>t> QszPekgk[jkm?"|fEI׽k/Y֛k֠zW_n򧡲/NGF/w}dicƳ(_zLҍުͿSG"bt5`U&P7;QOz F{G C9o/,HC](gDhp{TQMlH&v}u|ol٩{`S7 :ָU= I`Hs!:"t5d6=^4V))LjzZP{tC V%XQpƀ%xo͍dmpaL&H>' 8ݺsiFJ~B<!cRIUS|Hȹݝ6tN:q}I+*G@l LW-;ʋ❒I*mH:28"P-3tΘȄ 6# 4w^jipgV/G'i"  `nɧxjIGTn/^$Yg*#\vLi’`gW޳$AHJȉ5c=댜-lvaˬpZB َf5%I̹Z \{R=Q?{4I^T"N?3y,z? ܇`"Dd K#I $$yYjujp6mL'=ߛes+u}EOh, 3d$aT>8[ G'c6:Yh{,CQ`:(ATbx.I@ ,@uQLP2;̈ mfG%2z/יmaxFq8^Nx8eMi!s^`\=A6`e գ;X8c}LBvX;[LOе.=ZUi(#a's@B6y* eqIDqp~|z:p7goؗg>U;Z”TkyU|~4<GEBjW5)]^Fn*]{yE aTǜ>u&E}mǧ xY*g Ud7.cP%b2FZp&B!N4E&'y5e4M%EHNs\b!p0*Q2K0򣄢/rCz&$Q>+>/d}> ".,DG` T[gdʝ:6iOR뤔@ ,@"Qyх\@p\!푎E|+>}_lu|8n>o>VV},ҁ-9A.qd˦QѪTNx"݋v H#—~(&Y4BYRZR1PT\VeǍݧAd5mNj Nۗ`XP&sJa* d<BwIvp!M3W.iq9}*GJވQa -:ތ̙;>,gnCvY(DHE;c"x'PHGu̺ǃ[] .`(`|USlPm.⨘a*\Jb;(էqCS!˳5oޏ>Ɨ ȳL2~7wz O1M]Ȗaloy˝ ' tdr[kmp3,y=̧05,? KI*A1_L늞b= fb<;]M(sPE9 ӳ׊?`^LdT|h@!jV%f`d瀅Fx `t' Z"+'4Z0+a1Xej76ql#dpa*^U)A:!Ҕ':ȶCdwB|ee"S: )ɜ_n{QnnCџ<. ؄ SI)m`*Qfh-H! X.&rﶳ@mg!W˩ w5.d9~L%|s6&5syaǤ/Nmr1H x 9^oI8=wdN u$XRlNgr[=?{B=S)xBeG,~QJpսƗRRYme"BYtJH2jNYfV[^5=274g#w \&j t"$P4!fb823~0};7H{t̪9E푓8͑*ts Mi,ߺvpqq#짉Ar33?;kMrb=T2G`ڻ`ne 9\my1A,X.Z^:l(a },rOE0`z-!ߞE6\(APRM3D%VZ$|BD .jּ!ϙلiGsֳcȹ=C8e3^B 7X ='hj *YzZƧM?e[7)K!GwuAJIb)oz ot(olK3*DZ"v[1+^JRfۧc&c 7(e+\ #5;;gZRHR)2%SQZ&gS{Pv;,,jm9\hL,Dk,1k! &Q…: c^-J*\~ ftx6!z!PSU5_ݭﭲn|ײyQNb rbF+|J(0AHcUn x2zӳYt*>D1PX0'PM-?;-O''ܟ\)+? [gc\QC=L411&bRʹ6:C  ǴTh#zL~QJYW] :х㐝ۈSud7p ߁Ȏ.xV.r:쥇m6x i#оS* ; 2Ɛ\tlbiU;MT< VoXw(jpu>ɎC&x:Clh"@ Ϊ#i5*J5?C`MlSJr\+n]6- ]ԕ;n'l]EoR*2Y?O֥:.~ I>`/3joo#RwϪ/)bO jiEoۛGs&[J|Ŏmz*mٻFWb)Vͩ=x%9b#݇^+b Fl8@&Z!qFD:ȯŤ>x(#w $4׺(-cIs|ҩ7|=z*'imbb`O{ϻ8D&]zKlDض}pSSh=U0߫)(y%YIMiWD#]pwF<'韁"C hUZ'.N|}vg2Y9}&RD{(b,\n0 Th]0In<,pꙻc3V#&t<&j^E+žza9g'ѥ'"r9z;Ωvz2w)wS,\>J- pT*ҥ1Β#&#&bQ#,fyۋDpspӉQ1mHSIdZ X1(TY8 tS13o~;X3|Ɔzӳli9 ckym*Z嬥pGۓWʛ&R=1M:ԇbjsut2,NpgI;#+8DЛbE .XSMz _oCK]8K@a0$0,t>PF7aKn껶~7wxh8gǹ3'aYt{TM[nԁ4ktޫ^I*Ku2p"XvN"Wu2pVW*)b}e+ PD]n6[ 6z F u+\+ƯEKQ!U:Dn!Q"WA1 "rtH/uq6oC`RxyBYu2LwZ D"#cԀF*%"o#9 +T=2;a-ﳑ0LplSCgÇ~= \5Yܨ|ڬ۬o??``M/`~Wkn)1)Q7  zahF]̋>ʎ5I/`¨ X 3^9{:Uvڦs$Zi'Ts)H0 RF|H˩ H5ʞ[e-䫣؍#n)Ŝ.VĬDhF H=4,6"aVr8ǖ#GV IM ]QT|Wc&hSaaR"rEaF ew146Nm!vNCllv[es42pnAkdQ9&cxtHHHQ#T1+2 %ZK@y@Op).}Ԗx93072f#02Uaaq*X(z,xG{sw$,?lhh{f4<GlQR(-6S8r`H0)4F*9BK/m&U42XOP` yIt;t$LEt!"'UfFaĶmAlTԖQ[K bPQ;$,PJy9MZ@:(3N 'H +(>rRAsK~-|*3"{D[dCXSh9 hV"x NnF![B䩕 8XG!1p&(VA`Ir4i$53Gc19 zj 8_E.֙픜:3.{\+ `i@!N ֒GA{%oNin &}P%dt̝(z\p5rbq"#>ra5xx 7 sG?~ԗ~dh=J4Lgl|!ޑ )}BL)0H1ʭt:bx ;(֥((V;h\9ՂĘT1mi$S"Ya5BVPH)+@GC*⍲8`1we-"rgq_ Dt.r-qM'$ӕVϺpXxOreо/-a>I'*K̲ P]WWBz GR[(j1bS8*hW}D9yO<ȌC!R\HԔw$ŏEV xHgֻ0pf+MI^ӻ{P7W-+˺;{l\V[ﳞ?`ܶ"f'%FJMv82$B "&HK.yF) 3_U{5.OnQ_oMmcjÄ3.ֶ~Zd~8]$&f@Fc~A[(@|3XR?-m^[=II͙xԳg_%ohb(bg!8O;b+wyR>JqTVaYT -J)@%RHDcni.9Mz>ѝS;1g;ghɂ.R~ +>v+,;eLR/c{^+2aq0dV%WU we,e1⓿rR"` .,,beP6:LOLj9+j$:yfɝ' I}''1cxϛ$ 3}^}ޖղ{^6{u4SF(edTLgHՊcoY6B`$Q)q}7\Kə`=ekgA\:SRu<:DD"'8_G*%V"yYBr ?%>K"YW^ QQ!/P%Srպы$b2Fjd8瓱~s8I+2,~ ̰nJVrۗ0(8OXw?7~,κsq_0daBE ~_QX}\u_/?6yEtW}9b'~yr db S-0<Ȼ~]eZO9ST=*x "LS鵠 կS<-Jr&_-%_'K0+٣'Ĩƶ 2O| ^(ͅLo 7bm ~Y!j|v8]OymrqZh{lf | hg-US9. piԅ<5rP aJEzջW[z;$1ai\AFJGa) \JKBr8zZ()Sk:~x>rE0pkm+|^.i3_o_.U+Hl E4R 'aA&PDigp!8:ٌ[:y-Xh] ;6`#*Bw`P9[7ǩ{a={[ܮS, n+z+Yњt(LHq"'jl{HqR1ԂCpJUJ5Jhs Lt\t pup%KpIW\ʻWZ*JzBυ+ J(Z!WJEzJ\=ݧ5@=rë:E0Z(  D~&VsS}$S4 F7V5O>ys&Pb8*'>$?ScXYؘd^6f j[*aJ9+p:VrT,F Jly @0Qs3VD-k+Q)X/7\ܠBKpcL:WP{qp|=\]\iEr HwvU&piJM{DJ0W&WXjy/WGSITS5QVpuc+ 0JUePIQWW3Q,ڲ'~ym9vJ-T%,"n:xQE[MKiyD5 X@{\Z"WjDS~2QIE/]FK0 X`JRJ vJTJU"U"UVUR87a>¦_\!Dy[qZpjiAX3((dQdn>,p*ܤES,e8le9SQeM(&eO^hQJ' 5 Zǿn hC5d6Q{4'5YIXb$ȅ(R*ٓ֎ͯ@?OیJMb)뚯lPMdzFjJV_*c;(ݑ~6 M2"A_ 6IY7@+fmCo4jGh8rWvgmh*X }j:v6*{3mjګ:sF (܃Y +H(򅖜N)bxЄ^xǦwWA@s\0sX̦Xۼ&ܻy~h>Ng!L>鄏 W3ܒL^L37nE{޿f6WLWiOV5AN8֠.05\գ+n'i:(?d߁km|7$";w0cd=yHl*Ntl IIBަF4 4FޯXMk4y_Ud;@rQ`L"lcIO21!lcr oMQ1wPa-z vJ∥T%8bBee2s"pq,[QjL sێWQ8(3bLŔ]s׍mDzwygd{lc] +$\8m/0lLfѸokY\C7G̀fμKkLؓ.Hib`Nbi*qa#vX>q]5Bn>Y -Z>KJVd}pX椱y4Ch<<hRa^z4\gC EVu~0Q;g#WPaU3ַjkZ7В\5[r :}rE1\-\䪏r9he"+6D#Wˣ+=/+TjʕۘARG#W\,rEv]RJ7UJr]5/WlE4r%a`c++䪗rFD$W,#]!bфډrEz]Qd]!0xbW+]]חQvA^MqcK Bɚ7j,L;e.nR3/8@9R9 _4N׮sRhmuDv)݀{b֋5CFXnL+c1Ee!\ȕ?(Qq#+h艶Rrl,z>f0~0Hײ\55-ۈ3ݎ\5+@ W=(8x qŻh\\PN+ŘF%hۚQԌRA(WƲjտeyk &xc}CC]kbG CeI~48ph}پ _\^7J|5*-ոŻ3]+$;8pޒlt;2q=|Kuf4?c\ eb2X/w#\.vhS]7S%&yDr+m{i]3ڶ6.hF W=+u:"RZGh iŹY\R!nGZ30r%b+=/+>`>ʕqyWB#'t;)$Һ6derE=j0EOa`D4+2;~S#12G6Pc9\A5ଉȓ3&'G"Oh'GjGO:~) +0J#J\PȈ Pɕ3& )ArȕYb̌S̀u۫ږx6瞑vHQBxr%:rEDrED#W+"WD+e(5 rCV \!TᶽL#ɎQ\ BDrhp!gh\Ѓ\P2"4'UD#WT,rވv;r06&g-c]n#h\>ʕ6NVto17L4Z@ N m#Ri\F҄+a ZQ4JQ(=&wR6ÚNנB:TLAaO?DZ!; 4CPc\j\!6"\-b+ur]d͢c4lyr Wl]5-E'Q- JrulsɈ䊀UaDrRhpEVˮQvA^EGg0PYFj]+4C쪏r:rE.*q+=ȗQ!vK eD ʠ~xyItb99Y* %x>cb cјj!@ uS(n0^Tf(F$Wl8FWD#Z.WDٵmzRV\Z ]+uGIS$UQn"(`PƘkU0p4x`m{ofH l Gc*~L[Xyfw1Ԭ QMҥ@l5 [ M12T&h (\ՉFau=swpQ/.oW>eE"3u~]m⏷oˉ'7wyze­=,MRꌥu~I'?_vnnwu ei#ڻ\ of@?=.)x7<{r(nK zjzgiNyTFJ9Oq$er|_~,C5Swflq'B +K6fEXY !˹ )x S- XL;l}[w!ûyvc`\P>!τ!gR)s]{qA ZN^tJV^K(VxYDϾ])GLFc!⃮V%JJИ_8#%>߭ޅu>(&կFWh)] S1>2ը_`"b6'0NrɩMc[Cuz5$)exSiܛFWWYImd뗪'~Jsw St!*˽2E.у R<`. P+.FASQT֊) Z@zv-tot*^|FKOh:^  = 9iݘɥ?ܛQl~6Ubs{,ƒ5niXۆmx#VT5%ka{ec 'EǦ a9L=yK roX˯ nmYWLL=EˌOneY )fRrHN`*ƪ@D sh nSi`_ӰPvͻC O̱}Ff[,Bei<(y5P0^(aɜu)NC$?TBw)⧎D2A淋(1S>1h;Wz>s D{"G{W-F-?hb^[Ol4+Odb d=D}?͚OyaD숴K\LO,S?|\aPu0?g 0a%6(hZ,c`Zg Ў务ƅNIO.y*rFBBM '$p羼]\@vWh=>=)Jb,[,6E"ʂpʂhCCǵ'm2-?u ~Zhr_/˃';tvYrya6h+f'8! 泔seX!3J RSF]`(r?+ƗW:~d;Q/s=UB7:pHE&Ff0uRe-"xlA E-H k 0Kj1NXmۍN$SX63 ps©z6 T)e ~TZW>j5$!rlO&hضЃJ-.fԄ\k-*Efl!!Z(S Ҡ3 CEs'Ru -hB{`Wj녦-K2vvk;=awFaMrK ]k@dFR!F[(K~fV7awmI_!SrD~ 0p{ CpF?-)!))_3C8E5EJeΰWheh 냟8O_Nȴ:? 狏]@oS6T2~)Y Uu(4WsS;Eu^-#0bTNG͛hOmA{}{WӳՏ*E(@.7}kKD-2+?ʉ^6c|!8}K*>= |9|r&5jaHBZP!Tx!X BPԙ!GgN.MPC:,K,QpF%xoӊ, #`H& pN9/(}6N"Ѹ!,p N+|ĚbyezB'E,{_)vfX_`пʳ̈́,CR)a(LJN"ʸeF,҄qgLdळ<`E"%.Y? BM?* Q8hЎGHJq$'r6MRL8Y,W3#\TrL*04.gIh :-X Y(g'W̖-OϝR0\Yp Ht!Yˎj5%I̾Z>D%^nAO,[6ĊР3sQ)Ɵ$5ĢFU2ΣzQDqaY1x'e`M;c EI6 pTW Ay7fp]]SJ8n̂[ z~*WENMv<:|Dj,!^{IYx p+hqOEZCz('w`HG1W:taߢj7VoGgnq!o[CJ $r2 69R HysR:>/4 8ZD MFWY ?RH>yrUp Ua}HJ q T!Nj@!FAWH׌8hd`>YfEIKnTO (@G"\ŨӑGCLST>* <,$8c.8O.z)BD?$}y a^DHȾ^9124Zf9(&ΈI[9ʽ媯s2:)O AuBM p@"K>P$J"B8J }J=) ,uՎ kt26Πx$w墄2)]?frڗ~r'AD*ι4R @m6: R㵧Ga);3 qp>e##}S)Ւx>}fkگ[ DZ5  4qR$с>ipTȬ{*[|p\rdopq"+(:A> n.xy;BqC[-ߵcu<wo*i6#G~HZA/,+l'h4+ |k֭nmhvmr^jݖݰF3v +&g] *4 a>:u uQ/ǃ \GhFN eW-{^^I6-ǹbGh#aPJk8&ެI8;8iX'ڽ`v̀h{H&aj@L{<\J{+Mֽ9(orf5"D-јthXĿ1-FhȐ(8`zGγM锠*/c4tqX޷= xuȟvRopą]L:z> `fY3(E#0S58X2/^?^1|;!Z⼵!1A30*e@]$\[#d(;Ҋ8.S4~ ^eLdrպNuV^5ft$Ir1rbrbOm.;'S`6B F۬R+St)x&EUO5{<%Dc>'"5hiF0 TQȼ^<[?|AMy8Ύ9޼sqs*{hbb bBŔ2:C%cD @/Ť$$º-:dvWQ8d0ԠJ[vC<d!TBgnom>_⽛uN)F xa)sϝ&"j!2lb)U{}c[2p;1m_/8>4Pu,^Nmn>||ɔ\~/ǭ0? ߮ȳ.6|unMMK< ; (m۞E}EuRܑL/&zZ8/9,Puڦ=iȳz+ՏV4z;aq";=o:߉:I]B]M+-.ϋtZ[.BP2!aLK(5wPP{qb'֨"ܯz{U+3^i?}"nuw_1j| YSEM`Ӻu5l>X̷#[Jy%9V.R;ǃ#"=~np;#fv~d2tBڢ[Qqclox;7TWKj21s}BknQkVy۩m49ܛ.`:^`x!0w0]6N蘠l𞰛37l5lh- 5gNYˊٟQq&43g70mf$}R8p@lZzFtiw,$E&~?[}~bC:D+ (×v`Qѐ" K`Ѡ9 Iu4AQĆ=^Qz}RjBi?]7dgjӣli9$Tą'NGvQ?0rcR7}iʛn݊tdDZR:[J_.(Km5(❨|%I=Q\ "$\N;zq*TXc)[ 8ϴI k F'+& G\2Qi }&A2 Ojdg/o^1ޡEǝ)9 a9xi'5kG$9 R j;/3}aEe~lϷS-z$UHrnJ Z̦UEV'2J@>7a:v;xuyߺU.j^[i 0o1fzRVz6/TS;.=fg-1YOu6[R ]S#bUD7ZN&KIIH:tn5}!}fo۝eq7vqyӊew)7b7n @z*ɲ]=NL81e/ ,|H"#Ȯ\b kd׶ }H%ȄϗXZj/$>JdJqcbz[.qR_#V\ i.B0R2l?KS\$͌)iEݍ⯘ٟ}_5v:8jث=ګEKY[Wfm=I>Y?:\{j!\ʨ-a7Ѣh57Ϟsqê|8ڨAE\ӺwoQx!6ͅ5p58pkױNb<F\->^ʈ2#ډO趦m;G).VOQdnW[K-C'-eDNK$#¸jFT Mُ3ys[pҞ5d%T^W xɥC\-J^ʋie^L'^XQYSJP[/x`tZB>||gg>ic=wz=L㶔\|w~Moj__ msbdyS9HњE:=w_xpۧ>E;0ĜEBYM2*bCr-$J:$ZϪb]=OlNg+ԩ4=5Ail3 䫳U omzˇ^W5 v.xkȜ{~qsbss{upIƐbdX~~prw _3Y\oY•-v|a>Rzia6Y?Xqgv,qYxSZt#OT+FGx|X_, yxBxYgͧ 7R{/ď+vC w|iI={a5fNBB%q t*!R2er`ei_2{ć# py.?:zs~p9G?2UwEՕ%nE(; LcjL]/ɗ\RQBR!\΋ͥ4_[#̃Ys}_Ȗ"tc,4HyX[c5<|T,h( i!*93/c7k]ИOEZH)s^~j)unFN!IBH 9 biɌnGB4Cc1mB~Vgvp꺥QK.j8_ D4!9~pRu{ `> |Ct8=2U`-ѬԎFVIm QqQC{g}0s4ED!X=@#\|C}sk𖝩JJ:w̚Z! X2,JTcW̓OT+k1"YUI]/o>TsL FHF d祥oՑ)ٚ|X@{"Bd=pfqu~#CFs$kuTJ$ڻX/*غci ВkDªuZk*ڐ.v9hJsc-E.AVmYr4-D(u+@e s?Y {s ,ӢJz_nоkłtͥ9b688h3(r: t"'$ǀ{טxaS/s2O:{/~[E#U¨[&jm3,aCL=C /-Sn9@C9*-$. IWslr6]e 3i<Q"j.UBE {[ɐRa2Q,}V<.-eEsüX͊+>C 6(w2DhGx1puPuj'*N\gml2|Y qH_>ٽ?W} Wyg,j\g_}s6C@FGˈ34$@A!/V90(TNk]/M:38 H Hy3 ji*!&!Kٹu"D U X͒%z^bZQXyytL2БDcAxΪ$12nXSd~ȃ ao"G 7O 3+4\v@gp(!0c[;QS \jmOR/h?-Sdp<j›ڀJ5$sꭹ+ .B$XHĨI\H  |4\-XY%CzYZr,`b|;@כMnv겿8I0q `0uQ4t!>H`]=)F`gBwgW̢qѬ9ך-5ÜgBY-ÛAEF.$7*3x,LUU.HPQza2:}MUZ@z|BJyDHX[M3,cBr7?uţD)1|rp=.9F~:kZ ᅍs3O[T1@.>jTyhDbrT],~x|hSڸk㹏 <4gP8h "se6g kժU:6 |5}ΙtqCLyR'ZkVb][6B_w FޒClyNIApK}4Ъd-pڠ@2XAikEoàe BV3^ ~ iq0-z|dNUG 'a9C)z=C5|@<Ap.-U1jK ]koG+vi"]KAPOKkԐl'{E=hev7n:UuOwUtIJK1A&)A0"!2.bmw r Z56p^]puEBNeg,0BP&@vꅉ?:7dz X c_X|4 6!]MF\Ϳ1<4N:JM4Xs}晃ץ9&t :ho;a\?N @:*tDU(9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@{ȰGN ,DpeohQ:AN}tI"@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DNuah>9L@W8vwJ N c=&9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN rh@ '2zU@8 U@9 䀮@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN=rMN盩V:8x7TS,u^_V ww@S)y:ϻh+i#57%@lKKFƥc\rzwȼb6 |t\kf]GW4Bvt刮[Ԡd=+=s7tpe jNWҮ|ԔjJH67tp ]1w^]!Jp(y ;i{CW >]+DDWCW-Gt+kzCW P*ሮ0M7tp}op:]!JIj/ WE~[8<0 6av"_A~\3isL%<.HD#Ԁc-`24P9stEbX0It`.~t-r_NmAYi8ܶ1&|2Oi @|G@WyWW9A4|4=ږNJ>CHO*v;/4FMM·Ir֯ gA\2|3(!Z89^j-P^W>[SME4JxUABzRTzݟsu_/!ʕIL/X0[?/EK^G5RبM1?_jNB)F<U!065iJդo>STQn#dQoյ(|SSJ7QG*JD[q/z5]r<MB@BÉ'/]=+U Buar?!Ys2nt]xnVf(Eft剮[\8E τ ]!\BW3uB]!] )s=+ ]!\BWw^]}+)gGtM :R3yB]!])i`t ]\z3DvbDW{HWZz/4w ]!\Ҋ1!NWj/( ]+z3DtJ@tGt夷?d;`IP2i\.?OvB~Vf4mr5É4wJ`W&)i|my!6%(QlcM 6p`UG;fz0ju݀($ݰY m'DW{HW^c^ L"#]r3t(=M"aU!ӚfWuwBWKW;y7D qNWcSЖp=+/lך}?gj {feo z.BWykNWRWHW(L  e}+DE Pሮ])c;0վ`' ]\e}_ xjJ[]EmսQWv^]!JOsWHW*x L\e_ Ϋ+D>ҕ3~OJs~4&):w/ڰ0/ڬ\|k{}`d3$(d]]JSnl|Y~Bd&]hϜSe2.N*wA*)7rz*[L* Y't27O+{hMg}~:Ǖ=+KGӕs)Ittq+DuB7thw QJCWUb3 vtt:=)yp&S Kj:8 #xx|9j5. %2K50Fjƫ BčGa__arr0S«G<Ƭ9,g۠yukC G-<2i'1}\̄ՠŤ?@KvinU`. Q50Ь1rʵg_J$zX >8Z2G!U/of[>{;nn~].v^᫇pq_[ 0`a_n] ?\ٗ,pGaJo?YvrQ6~~§]r_0.mpo+3+i3>>: %fdI4)P 2Qʬt&|dyz?/\] Pp&ˣq:o.!k qLOEz@V}UTUR: x9ae`VJc8X/ǜ Ga%< yeF!.Y~0?Yý8ra9_j~c;=9k@^u o}xejK~n`OJ>,$I !jO%mAڋH/˛BT൯U ^ǃ7Pp| Dž=d/?VI$;WMBtjPe5xY3iU6OdE/e6 JؖT{;;X9- SXd-y t1|<'G/^x'?lA~BGdIYG8  3'O]YΤS܆ 4( tYƭ R9-cUY d(wFHVP"l.)aT]%TJʅ7"(QLb.|[~ iCy sON/ -l}]^M{n\,&\N,n?}'R?|cX`MHYP1q#xhѳ*d) CFOsp$ڂF7sn`Tp\5KUZ+NFhT558nЎ'A^mWFk+ Yhd)gr#˫JY)Og؄Ms)F Gd luP0lsQ&(Vg]&M>x~yڰ17zH%Ix"DcUSY #",[ó"D#s,4f2c"}"|ue2@3%M$^֔Ztl IIͧmKt!55~B|TZ]KAOŨpp@|:) Î J> n >8ǺͻhW4bE6mPцUqX?Ž#7(_n<gGp8&on2W ]xlD$v$|)țlo':Q05\TfJ0^=gvp.kZt'QX]zٻdc6kϼaU/d Z քtv\ʪAq}Ws9c'x9e0.,Ym/n ηgb&iB+Drw[O]nawoZ`pZ8RU:+;u0Vt"FV 3VNhky(4RY5%IRbr[5{Ֆ<09ҝy.'TUUs1"V+ā5dt΁ďQW 7_P\Q:z5]}]TX] s[NXs6X2s>Pu6Dـ _]խ6FZsq*̊) R$ H8~) }ReXؤetU춶IoAYuj g2DӁ;V9l+}+g:YZ}Q7iyV|X>Cqy WUHl2'3m."Diw;k,n7k.s8tnv]]%#s(!^+PI+Cp".65ؔrDM5+Y֖zdKy=?]%-ȰMipArֵ\v'S\UW4éE,/\_'gۮ}qQ*%&Bcd^DK'2%Cޕ6W)/{5z<c` H5)R&)i4}wGV/K[]+UEdAEo3 >ެIpv>s:z4;2+0fl b&/ +˖]?S8j+~Z@ʮӾKOž&鼥#vA])mpƕV A qsF3 l}`@At<($ Q,M̡HBhdr1)@ !.XwMӴ-,ڌ7BmLI9˴1o-9'g)$I-%+I!EܶU L+`IJAhô:iCX!CTDCAe- 3w<ֱ(@,?Ye*0Vik9Frdg\0ZsYh)zC-5@41Z7 ̵(ZS]?7$(!BrY1Ո"C?('06/0/2O@߳WO}v-E\~?g6 ~gD.i|ه9f,̗򲒀J%混"u>?oQ}Do6>TB @yЬ^~NYT\ iv-脔osx6+Qw|㌱ۯ=wߌbRʅBo>>"w&[}k|OR~v{==Y9Qbo9y:> X8ʏ.zI*O?U~nU"lba9ƻ6\ ۍL.X7ʱ@5%soum|+DR,wi>9Vz*-5xTty2?nW-ʜEt?]M\t;,p5~EO?~l*Orz6vE #~w=nׄtӽ ^ BM%~)[h>&ze~1:r+,u8/- %}{iy6_TmM0Aq)F *VY(D hd;!ih0>d0tYIϯ@ QRyxl) X<%D fʩ,JK3E2gTQMav{Lm VjE&)gFvWaVX+vrD hKD4xi RRԙ!GgשSQNՇ[9Xd&K %f49D%,…u2(5b@(4psxA A!Dre88xKY ,^ /Ҝ n=9\֡gdp҉e Ak4Vyy)|HRircΞ" "e1k`^¸A;Q"ۙ䖏S_rvD ,d*d.Mh#$%zkrCR Jz S9s$9g*GbyjLi’`$B`IAL1 F)|-gX1{ 0*^>?wJF&s9skA.;Քh }\h:'r@u4"Хܽoz#gghU!F'a(.M2L>Hz 9NzYvއe!p'ub 47*fTI0!8ZR8:QguXOòv`0ɁGe r\AD18 M}#0 ݕsX܄o+3DLP[5m1vM%xt' 0VùHGgrt|NiX^kᐲR6C^CJ$r< 6y`֝FU|QFjv=vD8R2 ^ #Jx$dk &Ħ '?;m|F)ݹ,r_;8t6ƙ3ԅD04"J2ex̀QĀHj#6EUNA09WE63S up7)l>n;4ٕ[gC"3;azXeG5H΃ tf?a;3(@05xmPSIъK'/LW _b\L F0LXɅIIzDN"qbAH ,yiLIǐ.OR,9xu:#g>s4~4|_@nlۧN^-ly]=w20}Gy!HR U \@]΍T)2",#FUԩUxQ앫KAq& 7dyF'B9.+UoR@1FuWD{I @ g5iV>L `"v9'n{GI~\u֪}}tP \IB))8j*{uEZ!bowY۔JlZGdՔM p- ]P67Ǿ2Tr1era6<%`P>f?\v&pC-s+s/H==f_+[@;ǭ~ziۍ:;vS챸.fFߏ^#gWnV ë"mD ZL[fДGqr7Qیb́W5Vf1(rCm֚#C=ZhشnT8#srB nhn4gѕG5uے Y)oG7n)'&1VaP}+wώFs9ݓf2kDme <1Ia4A0ԌދF;&1q?mfnZVSd#vhn4778wuhԙ.|XI[i\]lxAA`cXNZ/ . rs!$2IEK::Ae==bs::+P<]#aՔ0!$!'K)q$ O-X|:MDe <*ƑȫiE]"?WJu/%39/l Ѷ;\iJDg?Y%nDD.En D&P& qe"fz]qi GcBM܃ Qy "$"0~6댜=)s2o;şQs 93V;nnlx:l` e8ss # tԜ!f!Ɠ*>mBn*._xSM(yW|?&|x<=9#>NyJy*uZjE4GKD32ס_*JΪnN|zsYc7k~PFh& mkWwP6@Ynnh.&v'Gz o 5ZT;hZ+KA>wqn'U"7s/p"f̋rc0,HP/OǛ?zYNf̭!?-&w`A#, Qhˁ̕`emޖ{Y[CiAܞOIer,DAITzKdhhT2{O}'zԉ ÷uaGןO[c RZ6< QJu$sE}E7;,7uPbVAbü G5#aqFHC Jlk@jVCT6B!^)?M&Xҹɰ.PCc!(zv!܅Y{"$RQ 0$q,i"(LT"ĉR*q`*Amdp$Gbo^Pcp[*P(O!ʝ;vuFΞ5k*\C[C9nh*=.WEv} wdYUNV՜J6쎩]x7j>Yxqg#"Vmԕ7.ݹR-Bo{R<7.fz|6|n{Of^|救a2oo~uwfx{xk: Ҩczd[P/KV Ib웉;SdP ]WT"'x t+q(Ͼ!'&yl6snޝj2hWsfifmLi^nz,+ll_%0}%[ oufWrg%k @~/ YO`E Ke(lsNGP> E;ĝك$z;"K(Sv!*`IEਨ.H*-ybc["@4EP@.o/LR1($ SHpl3qL9>vgR=w<[7ҧ_7';s49n!CA3rAZPWdU;`t%_&EXI+5i=+V j]t_w}!ՁDsD}Ϊ>>gUZiJ>gY/sLmY`-UWѱUJ#\FGWU ]:\Uqt:9tb)^!\1*0*;biUR^#\x W n?'7אWkɳ6uO8)|Г䗟nr>XiN~T3?Uޢ1R{m@yLkv,*h쪴OxRZ5+i+Qk|q:qY}_m'EM*yZrL< x9͠uX\}IPޟwY)BP~7b/> [?JQ}9o?~e >t9ԁʶ DWR$p,"Q#?YťYDfi"rRnk?Z#*Ux3E*WWjW_?׮ Uə?_T䃽ojʇW'5jn[Xٯ됲>\|N5^t*R]瀠F',i>Ϳ+o)8*%g< p BQmyEyec=Fpԏ~KWRo%jq77V1_NC&?Ňj/=c7/eb 7_G/ח@fP^7zb/|'ؽZuJ,/C&ȿzW#?3UtOfi^I2' QiiECʪ-R,BxH߻l͚m%]V^d'g俗g v^2ܗ|;5?eq6mYe:;Գ;b+~/wzO/n$n|`}P.Tl]~*J6=/jKv貊>S阡+CS=0Cw{oA ]-ܥP YkTLPLD #D 1*VT2FƘ )PƦ߭~`>urYjҫYxK`-NE;tGZe1X*Z;>=[8g{.zuj3Ӱi3;>sfz_]/Zgև|GwW>:W湺=^E^W-zjcFG5_5u.tWf Ћ?b't~:_NO 3#'rG63 l}ni.(e(u&ٶ۸&]Cu|RBdKZ.3%"ƈL;gi j7ӎCQ6Fm;`wD.{WNYkC(1yCF0HdtE  xXQ 13^t2֤,QYQf RT Q7f܏S3Sq(L?EDj4"∈4>RWjtH!+gk;Q 9yRdC>;Jq"ݖp$3f-]1VdA35F lՙbdOZJ*DhN8#̯:0..֑KCZl%knw2i=9*:KqbAidZ »14[" e\29q)pq06wD#2mdښh:/"h 6"(ޑhHJQA;zʺh4# }\(٢K} &8!ɶ5<_%+$խxOAfؗsȠ4&&l < WADNLZVOW<$/>9xnûSuR͔;_m㚕>7L3ocN zOfYacČ|TsV~ '"bYQnr(HdȢxA=U%)y0Rۤ"IpTTĀQjQRE$˖uHa-UNP"( X 7O&Nbp׀LO:KJ0ט6gĝcwv>!ev4;߭-)z͗٭v[||@PD*̮\2Y@ 0D˄"hjB&63fČ)'}X I+^iD)jUU\)@B<@R+P{j>Rzoxwgu_t~0μ3`q-YxDܸ=G~I&gE(d)LZ*NQsDiM,js[s)PDbT" ŢB` ]Z;ĹB!x`<ۣ ~nԝ"`y)@ItH$ٻ6r$zB_En8,fvo |c_)#$[o[emɞ6[-6*~_X5L'Z)PJߧ)HR,5x_C"5p %> N~l 41~; Lܮ5gQ!pzuZa|HT9/1X~JcnD^Q7)ZԅW 'ӄ7U{ϓI8)jB1V@IH҈ AX/UI؊'^ǀzZzމۯݾl3|m#Y(0&\]+n{to+1,F_\p6 L l̼ɧY'eY+oQҔ1xK^J A$oRK$ &\9 T%v֝5x+zz\  rًկ=ޕi;eq+zfyf vB2G?>8#i>3 IG>%˂?8dj Y u"); NhP}*f4%bI8$Na1)`c|FlDViRl&0rɒvI"<ړ lܺ&iW;ʴ׏K:xcZ6Z W,܇ͰypG(*"NF+3Fz<R2ߣsd/,OXXnfLٜY]*I;+&҅ұFP`E L;'ѹc+P5hvkqvxZʊV}_j2-܄ |(!)^\P17gjk yWwqP(?:J?Ҵ"!KJf'GH-fP\*XkGZ`I OH*!ir>-9 FWQ1DH2 noO _(a=^C~| ifNf?r@kNv'$|nL%\ / 1([..ר'h\iEIO}7;ng螠g^]L/ߵ1Xh1>B_p ̪:60Հ,{uvj,[Ӊ=Rj3ggy-l})c/6|(5 boK)0VQ]Rf[z_r8lj6MџEƽc?gg̳KRi㴫aߥ}nH_7b+2|>k?''# ~?5Ǔ?-~p1^o\ r[fL/>%o=+3X n\Ꮳ񗷷ژϜnӟNrWV|8NCi'}ioT{YsKI8m!fXQފm<,I΃x9!bRŷ:6ho N wubCnEp.[WbBl[;Um|ƣOgg;]՚Yx \[$$*WfgҾ|Y3<2h=3Գkjxgk̋» %l6ۭV[/3+1w+zwpkd|__f8_C*l;ަ'*8_eMC=O6 ko b$K  ʄsv\fi):P!. RLKQ arXkUoѕ٘"'f)FEI{TQMH&,?t1;U'4 tk<XՇ!' yì\V(H FAMA S};aO`Cvl[(b^Btl+ f"lc$$&89]bഖ j['R#a-4kV^ҞA4:`#;p"#Jty}(+1!L]*w⃜XB.XZ%Hg *G&&:Z@9TffFb0ۙ.J> M9i\1XXPF_3$GQO{) $֖ZtzD ʁ[#"D"X'h=/M A-Ԙ6g';T;?>?w)g0-s fU_m@Rʥ=I'6Q]kNY!&PJ6sV$K2H4$ fqCΠVҦyqioqhƊIȌU/m̨5S, )34ٺ>qRe+, >tjX%ˆLGMj\ARFd5/x1R ˷/CYBzk/;a4QiSkq9L}Sfw3fAӹmzHo-ۅQZP AV`U-Z)ѧ 9nI^8=z{ʽ`vl=9q^kx2gb K1ea!Kb0M$ Mh4䔄dDW2"ׁx ]H=,돟^{7e AvSz/ύ'˧2׹QűGGGBG;d  wEi4夠-CQgݹMOtht.'ljDQ yvt{oܘf nw)9SY%"xkPZIR.;QTϣm fY|IatvsJmZҰ#- `9[mooLq:7,ZDB95qm (7lOu"Y0zHa EL!10![u`!,x|Hđ69p'd;;q5N|:ߣ^Cmn{k;nf#?y\/ب(BMcٔH%|-ZdylHg!-*-dPc'7=;MOoy^aPvVY>HE+aAPL|g>-3)ZCoy:<4r!2ol9^+89a<M9D/eIjuiIf;ouZ}G: P5woNv0U7m/6tx *d")^dIt 1aIL3>+~`)eX 'jyx( !s}

ىGcEpY8#ȝ$6Q$á,S=3$lHiw驮uUuuNف.ԕ\5ӂDu:0*by^j4DkJqk(HLvヤNRs"[zsU ֗k̴gO~\-aڤcXݩ>p U.OYCR1hl!ANY\2s?`~l@$y1"5@Nzb *w* RUfԿ<N<4 ظP`މ4wbGi*` nڍe<ŧ=?3'NQ?iY(SdP\=T*|g^F a˘vcyWά: Rv5&IdNlIgP)O (=xث3yb:1`yNte?zzck0XiJ{?S>go <+; ZX1mH Bp"p\Z`4 ,AkO5Fq@/ % t(q|$撍%,4}ԇqUf1Y22gUg\Io>kc<3RVd HN<F㿪ϙ' w&W61sUU^jp)S$q ױ 'e/#X?'$Ǯ_qi'8KY؟a]T DnjU[;xs*wj^iގbtcR|kqY}yꍆamr4ЙMSQJe,:uqQkMi֪4_2^Pp.`9VL'A>aUY=|+& _x_o2.Dꗟxc`\b҅Vpb#Vu,kLY1#|;kaV)EВ"E4a7Tv|I^R+R[0EEg0;˜ʒKܞ^RΒ{bHc~Ap J슋9Kܪ?K\tpnJ3%%5)J2}pq5"\F7YIV _dJn-z? K'}r ]>XJH됽ts*Z)-bUkJdxhO".׫cMl\3WuOSh4g=?Q=^s!u>Ȟl(Y!d]_)7sT;rԂǮ~e`/n)}hl%Z@P h<\PMA9hCM9v7Ә0*bRc6(SX8S QQB!Q0BԤ叫N9z](9Ix-PV# ]bn51xqe=ɒ0q(c"Wq΋nH-?|m$_/Oq R{D΂/sN(8P%$z@Vg7Peb(yoMe4Ink0n,iKV o^Ǿ1β7;mr3+왟'# q0X A:G+U>V;v>ohN}4nw[}-%F&"L[$@5` b4CuĈ2IʪF,"ݚvY+냉ԃ͈hA #!Xy$RTY;n =5/զ}]{gnWʦ_xŨPqJw҉pk(D+TcdE$",7bC ȬG#A#(´`;Fn6 W* h$uHHY dVjMWi1 !b'J0Eڃyd"zb;MDFS^txFg-t֐zgL!.|^ (0|{H@Nx? Y!' $uݸAvmYWF_lo5F{C$H d"(H0u>۪Xv/;B̥/񴀛F/I0}0~嚧qxǺq`㦣ɗꢻ@+cD8**ƭG_WWA巽8=U -nFaM¤ϳ ,rl\NFh+kOFp?MݦC.ꂵv>pلXa@f~Sɀfp? /@2Ŧ, C/ N@ð3M"&)CGed$OW𛇰@so)R e1EDU"nL c!\@I[T+<PLFo' 7BXB+@akbKFb™t"ձ Ks`I&9bˌ١.m$IO3);`.6sBtyl֐@g'0r([)K_|gm.ަ W{v.@Hg"fFZ;)/+<+1Fe KdC37kr(i=-{]5/ukUۣk_o خ0Ѣrj2o;VN4WY)|n k6U7Gw%q/$do55<UJ! c4c$y\8od Őt֢}Rj$KoK*TA=GYxOdu6U)6~ҵv}k<|ɾ3,xcn%8@{>QAR$-)PeTJ8;wMA(9]%rkU1 o hq@KXR"d)Hd3QF$=Ft\,#(dhV8Ee_:$wIw/SzקW3r:ߗLmve:@PmP/NM1Ƀل&:E0qR7G*h&EV5X^pОihL.($1x6\iYbc9Dւ J3:N~$- 4~*k,R]k2lA:w|  &:F];*x(Fh{2J`CÈVsl߻𖦹[\9޶|Sw]zn6ZfBԥu0mKP2Әb@ky{k-f9u|d23A"N*% KWFHqXx K|ן]lVakb./67ݽ1'-zYBZ$wKeހ9'I!ˤ8y9 uz AB8>q!ɑtB#1U: fމ~yk}캮hn߮a3 WxlVoStEH R+u"{#Oie煋!/.»]3sM0eLUHAThȳpdx50Mftv1,DP2Na#3>H B[K,TcQ*K'A0ٞVW%Me):ygY.IyL4M<1d(ef.x!QAlxZKU=+*0 }Em/̘ܺ41k4aUWNe@Ci]-qDrH]8*ssB/XT?X&NŬpzXDIZ:.4d{ 2)di-Z[ /< ߕj: њ$O6͉Ȃca}i$T9XH"Fꖑ) 2ȤEL9/mY'/2pZ` h`E/\7Biػ oqMФam4ۿk_C")|GiS15WD6<7CL]N.FK P_1%ϗ2ͳ46=iZ(}#{x5:^\gFDjD;\ׅLekAqmaEFuvVs@S,}=Rj3gg%﷥M MiQڬ>8lqz[$FlrnjRpd. GcRjEͼuFO7]fIi͹oٺl{Ke7i8w~H$E&7WcII['F/ۗj\5q6ruӫxLW例 4uE!Id@z"f4"Gg(\ *>>1aq?┖:~&vtl20i}QCn:vLMĿMOc"In9&ٹwݷQw<":`  bnYp _ѭutjnZkEJKT,'G٫vsZ',rC'|3w!I3$i_>nˋ88m3Eǵ'A=;_G_9S-@{z7iX֋A2Z?Y˘}崙}ٷYvԾc϶t- Cy6_469NjM 5{z &t =@&#ۥ 40|w<O];( }ArA9c:z[E6;/$@s@f<ć=N?;#鉎v&{|tG0{ C"SQvD0+; 9bLL:.bEH@V)Kq99>]bQbBl[9E&%+Q GBC4 ebRi$_YoO!2pB2OP͡&dxrJO:R 1"Y'=qN{k3lX0LתrO_ŌQ#YR [`Bz$b.Eo, ~8wqgϜA%LfE*Lv%za,Y @C&M"kP rYc|̡:΁|;ݱl\Ct/~#FKp[Œ Zv9R:6K7t:6 ,% |x, =ӿAF)R^:(CNV2B9R RCH,V:!J jI6mp{,-@V G(4/I1F'lqBc[SdApt"8zYM<Ͳ`="r"쒍:iGu& 56(R`iVΙ$#Bc+a%-/<9ADca%AF`1 dF8d{t+0VpGgr,w@ߣrQkpg~qփ蒪j/y (5(a'ˀ!䞂MAOSfl>lBR2(Jy*EHr…5S{ IU=#=( eS%^DaFw+A0b"d'H)N ,cttliŒ%LP0DrLd3Mg@&]w]<܌S8|i g(@ (n>!7ܨs+X QU.*l9 Ufbdp1e8*<J!|1H} FQD?F~B7;ռ(̄Ĝ.^مad(YDE.ȭ;HںQWC ~XoW\kΝZG%Q#K""AdH8g(I:)36XxlC}%4ہV)gA;܁R͔xV DioE)(Q3|҅8WcJ١_g͉~lxѼZJmb㋷iZr8IH^ߥZ>wp1ԃ:|kԅc"q<0# ĸYdU}7}م-cUz+[_Jyw[#iitbyQLvS O&n݊6[n{E۴ݫluOHK:A ^6m^f2bJ+ͳwJJd+U,jxud B^}%C0lFWj_ԓ5Ќ W-m{n_tղ :.ӽTDN.1#K/t,%LI}Z` vAǻ%'S85?[.@;0dh#غ_be.^|bXȎ5<H`[Q/{I&\t'2T:Mk>|+s<#f{ ɨWsj/?n'*n&D/ꑙ ~ȧ@Zl ge=>9k?ԐS"9IYHvqh{6nm5?-a<[;I ղv:npg~\O/X.>M\\ΧF F/J`ekTJګ{a:.Qr*$VJ# .+q]mo#7+\K|)Ea/f G;Hd&WlIdK$Fvf i46|0*l9=&ozp{0nI_w{:nDl!TzWyur4E7xe烱W 7G 99;Y3\ ˉ)IZ4/BrwTj/Vt3;!S)ZoV_ů'D*o/Tru$PB- &WqGӯOϨcNn~^l]Ug{kMNNvX_}>$?K]%FtEvE]ZM3JvUYg,pBWٶUA)UOWoҡ-۬B\$7mVپ݆(!&Rt+eWhv.(i Ҵ Da=_Ja^DZX\_旘]+%"I+YZS}pThQ?Zo+M_'gި!LJL#w=NcO?YsTm,\B}5%-F p}lXT}>MB^4e1BthW:,O1Od,ZIkȼ{aӺe#-J֯OC28KS6G)9nV7?׿OKl]sO ]W2Y~5~r}oGya'~7?>ƿ Kcۭlcaגh1i3Z\ֈiQd`)(K68'{gK%SP\uT Xc 1sMI8B|0T]O|Ѐ*8*UCr!,ogc("{J4f.n)xʅ>fWfO|fI,у`^#>KqMc+ߨBɬg%$z9C!y{bٴu5;Eg<9{|hb"dUW)M?\B!Ud򳗟l lEƅH2AJeY`'ڲPZHmS2B@PR}L.X:sbP@$2qLefy^~ i]y+4r]O 7nlE=? *޼]3󾴺pIoJ~'ý#[6g QPq(Z, lpIeIj7F=8XGNsMlU'3b]cPAKGNs {pcp:d<":_}D!#'>Il~+pPO*%|i&G*IH7 v՜y`:&H~=v &.Ф ||-?'-x/|Q:JҢ\A* eTAKƤVlg]8hT=gWBN@Z10ؼuV O8s1.8!i~!ml<5aݙ_&V`xv&Wx~/+mio-GlfLTIyKzTEI"Rh"?:Ցs&i澟lpFN396Qg $10h$ɚ)3$R\'#I,ݣ9cU_nw4f!z6'Aoe` /r#/@ҍYs(r>7HqʥlMdB_e9cs~ó 7QDi=sW)e.d&-OImKIQ=)){ay)) 0)`|ԑ+ˬfBw0jF`bS"d'QKRmtY: 䡱XOR0DrL3=NϐBxQoziSpשvZ̯Ǘy`4j^fCGJ`YuZWkHZ圧 Ch5YA9yN4iO4R`K4JjH6wKI(kHh,ѲMtM4x1e8QaHFdLEog3ف٧UiHAFKJ/T ݤP{鐘Secvs(qO#)"7[ ~UFU8q7WsRXY!FDE"IHqSƤXc-I iV}MD&U 4%`ġ#SP]S@5Y%80a!ͪ!}ellWj$>e61x>(om BF$Ueo- f 4{HrҺ2Krq5mʵ͟}{Io߉f |16Q *y+T#z aJy֒JGVj]x2`7z.ҌM{WQ\W{{W7aѢ&'ZrV*pDit٧@ cvI:n-OZo1y#k1}sFHqv_|d2zkC wAUUh8x2gwXuoZ)jL [)]):!$0b!fY8^xL1Y'flLشOzk&نnYAgq| l,dQ[)3s 4Ų(%z-<+C`ʑfl:at )-`UC9BR0E=_ܮ8K4JhκX(S'I1,rdwRۀΖ3))2V%כQ̿m[(D}GR_Q5snOM4!zM8>|?{yV5MoFg˿4\H-t.:>CD\~KT'#hK\1M_7}]M&qݍkJ jruH7!`On1 ?_~{ӀUROha,5k:˕N wܗ$'҄/Kf|4 ?u zp䬿~2\l힜/+tcW; nI[u+@K"nX(ΖmSSYj/zA~LY\6nՌ ~/v$7gK3_ gCŧMnlh/MC~ZOOРD/7@ҁ2 ķf27yjgV6:_P:Zǚ`J) L"]@ƓAЧy}'BK0;Jr91 8vCc7Y2: <*={$BFUI-&I CT(CσWWiU:nH o+ ,y3]YZOkR7XUpC#QBWΐ xQ$U@Tuqf+E S42ta`_e/u U>_dlii@&rb%(2GBh*!24(,wj! 4&hݜ)iM(vT*V<{PlPo@sCJQ:)<2tгz:yoȃ[WOjW|T|A},"d'PM-^GO"Nig;)<[QNygo~sqs\EC&1zfZ`؄D1m"ڈn'd6}YWņ=/(A2%ԳHQiR m PM:N9'm39HL&Ã4XQtT9$jc`)8C㊽zK{sJ7~:?a< w~&E1cI@?~G!4Pg|G&xxӻzW.g}dhBA)n&3iܥ>Z_~N0,vq5u l]Fx:b`8]_d9wOvk27>*U)xvs5$'QAm{TǛo#/k{Fs_?¦~4J5p_{W7u~y]2cw.éuRdUM*3];Yޞ/5Ndhfw64+ckENcL;(Y YC3}^ڳӞ ! ,\!|C*Fryɷ ^-h&|e_w0~il'GOJ'dA߆r4'8Χ4yBz6-aǮWӦc41$zGU>tH*Kw}9ޓh|uעzmxdB"5˹aV<vGg&y8i L.rݷ[)u^;e -K~mjYieζ<D4I&@hsh46(6Qm XSĔ>91!| }8+N@g4%)>$3A2B4B& 9M}N*\q`2p& /; WbSK~Nm%)PB/x/APMJ,Jp*aIv$kH9gyT"<)IRN=ruyo˸(nt{Ի?r['9u<62 LQYu>;mq_ᰳa>Y˃ &ζ [s|A-K-+tjQQWT@涏ϚxK%PN.!ɨwdM'N9WY^K6EHr^̒`YDP ^$NR+$S]O:Fo3THY"-:QY-R<Sm3U"tP3 'G;ٓGf[r}wP6nO΂uJ쎮SxigvJw>rhMcqS|sNlB1?2o#1)M#Ns`4.9c%3@$JWl%l.(62Aޑ4"M}7f(&KݖgF8~l&?bGXQ%o0GITdj3N .xq[keDiTpDDR!* +99uQdTS9!C&5FbFl7%TP8uajw v;s9ֹMRV"(& Ix$r#@:̐ Ԣ=&D " hg`\ y$H撉6 ɤ}Y"s< Bqi!!ڄp@2$!QH!M&3pg쬯FO=:lzs|~Vky3s!]"0';$o$场B`9q)A&yˈ$9J%)4g]IiY;Tiǂ[VQ>"v]_GwU J8(.{<>㠴ĩ'A)%]bo012Tū_e{so6Kj+>-s "6/!<_i5ƀQl{$xN9t )Q~nMr-ӫ~^7}~${x;p4gU==/7MfL[k3.[\Q3TJFGƙnm+4^UOiu92{~-sg/eNP{ /%r-:Mk4\kɇQ}haZ L{NrX5~iF~n;_75WL9 !)`[6jɂ8r4%` w2YN-x2$\̑DTh 8&;wN^v&!hˏ!k2 X5T84^4!Vhˤk=XnE{1FZ֗FUnGjyt ]]n}U;*pϚ);GNp.0z=VZ,6jLpd?na_MzţD?c[S.oot8U >]lXB͎[#6wYetfVZ\f]kFp]a~?Z/GL{92/*Rm@wZz=CY:,oTǽ wi 3a 0Zƈ>=ZQׁOzMYѳ9,r~.EZ}p~#8^H,w@q/B9UBLSNs!pM~/ @|3Z wns:xW'Bw$aO;ɏ @LpVb&>jc(,-cXtҳoUVpT=`<9\}ZzYƘ.}@4b8Qf"! "2!gWb8,{^*TLsKOJ(ELS[$T乥&DiY3M :ДSXA @{R qlZ)0q羟/EKcj QsWz(I>UO~\>_+pk'6CSk{fzfͯ8f;N_Yԉ.uS'ڞVD/QNcZf`q8Mq"#n++ .n)0G .M,M4@[SEa6yn|| tjC/X:eEn=yZ4N7g^ Jb9$ >Km̱,|*.\k& d-ĝ7~sec# pt6vZ!'P}WiyEن>YS%9S/dۡ >'z#{x/v[g~O>Zl2#jͽB5ˉ\ٰ6 q tw{؞(\YB=H Ce͢խ,U?li֠횕T">@br.2aN| ; >=pN OTFfRT>@Y SDn]v7]Zh( oBYiO7/Oӻ4zb|wf >IIoZ~xx+cC˭'wC9p}ȶ=0W4duƶ6_ȕh-kUUxLZ'$WR5% ^ ,pGHy*ϭ^RRA 5ՖS꨽TAUFda(ƞ 6AW5QgM`U%9Yʈ]M21մPԮz@ MT.yCtYo)((DU<QF̐ZA$X 'n.Ep>fv4&i*a5qacԗklf1` "VӏCTFD3  4>\wY4>q`Kѡ5It%}qI,Mu%a[lV3%F@RI1- -iZ8 kw+#b5q:\tp=|'_g5-9me\. 0*ZfdQS0V:Spa.ׄĬME .{kiǁx؎'gq w я/h3G?6[ Y^3zMpW>|(Cԑ %EQG#:Ohz{hzhzLYH圔* @I-JԫT&9u tvFfQ3%#fUy.dDDJr4Ex t-q.)=@vZyLNՉ%Sn5 ' 3NoI:5ث^@m3,|uz7i]*$F@ _]3ira@5 Kn%= =/&{GdT;DH y0 ۤӖe֡.Pd<&T[ $G]2HRH2l5q:&̼ v6OY:GUx}Mw:,ͷMlVs69v~;BL=wviYYJ'%~eFSj v&U1Ӑ 1N,Kpː}h%.$t׈g3 )"ы8=*Wn{?PmHbrGS^oD{L6˲*P'e]_vdu%6EYDs8.0ȫEt"H_G=sB3푷yz̭-Afw?8gM2eF=jcYB`R +%wg"&(w^!*'tAX"Ebuf`*CFT<Ƙ(Zi*rm6WMl) C]q2Gͼpoێ8WggoW{o_o ڼ 煨hM}p;Oi]hZ~2"FzggɇQ}tndZ⥷$sY lu@ڪX[ jΡ.K &!CtMbCUYb53E4`)S6EC]#"mոWP!ik~\aY>'>Keհ2uBTO:Ջ2`:A"3jg.@‘.RFs Nz'=w⒏DDr22YrVnXOypUq!BG *椔ޮ5m8fyl῏@1_bj%Q"0YUgSHm߫i EZˆ W޵q,ٿ2d]`^8bArGE*$%Y÷DJ5he؉]U}Nuuu,Vlq]AY+]*A-Sh4&-wlhrd3a%7>G1A!ȸ1^Hka Y֎%UpA |Sr٭p5qgY T'ۮO\6h(k{%3ot{/hB(4ysW9FǹvizmѓuPQ OJ#k;OikyX7Wǥ@hZbsvݗi3fq+:G 98+Zgs+Jm76` V.3H4y!Yp?ǾvL'S[< {Qc"@(#I*a2EOdU)kd%$(UtBYP>Cj, idd ` uBPt񜈑H"_7"qVg͐0[*O/y7?A.qLA4xRqM̚ūxGĢES٥}ޏEL聥Բ80ŀ,Zuzh4)F-eybgNOG5R miQJ/>4lco6z[N4Υ艿L)Cl -ӑX_T5Ӎ9~/~gW5tyvv6 4e5m oHhޔ^َϳ|&ϋ>TGnMc/j͸Ȓ'71KyWKSId~/+|5~zwc`_p C>Yw*(a})_TZ>1uW*P$$<Tëcpg:&hg  5bMnYpaȭ~}r+QL>Bn\Ϊ㲊smT^ͶYy5qGzg~>$ ? f;E{'yz6vE #~;gfޯ)w;KA鿙ZY_;֠y᷷_hY/Sc_,,mpB_|=vݷUVSg/m8?0W69Mk$W֗&;CHw)P+.JMrcJk筭dh}T*4Ch0^׾3y"OdRL&2w'a5DnMf4n c^Ȝ4s& jx_@fdUzZsz/<XOk%_ o E2] 6oF5ZsT5|ӌOс&`;.oV& ]N_;]=EiA44afuZ}?==-\M~:;C7`>PUs\=TKfJ550RTko︉^q^M'o+zhj!dG 'ǛHh? MfJѓ~%qg5 ˵ dH3w0v{9L1nzBP\¾/ gkffJD1EO=E2gB!md`z/ܜG611{G'cǫ7_]J íON/R'#ϽV m1*!#T!1c(p6`,Y8 :Cgb,_=0q eH,DSFt&A];gd^Y "Hq߿ pL{,X27cB|zóOkl]qcf[ZDg g%5<:HɲP2i2%\0WGK9xYݔY^gTVڤ;ӎ&3=[MOgy^!P4Zh>Y )1D|#>d=YKpO 3 o \lڿػ޸n$Wi"e2 v1|Yd -A-籿~[Rm=| u[nݾ,:*rrCz޽ƣs8߄;ubU ,>M^˶q;-lW}oʭjwSqߗ$Z߯w ީtQߝO!|F7۷7w(o}N_:v;6zwS7xu?;-M\lxQWSTW p._7<.//"|mezs \|M3e\p~lWX0[~VQq7OA۾![߁mFXJJi/7Wy-7@wkg+6 ~:GjW) {ܘ(sJYk V\홿~ z ӽ?Ola]Nq? -@o:I/*4?W]9qB/]F7OǬ ~aswD`ty{y/+88mwG 0ZzuXDm߼7s3\S6ywp1Yoу]ohۙwe' Xk7O]Bxnr }194BD*6@1'SC+Unlc>Hձ3R oL/fx˗F=]J Rް4LfF@U(PC*~bnf!n\~粍[ꁋݷӝ>Pzw> (WaF]x\u]D^Qah(K;< ̥-8N?OiǡsX#J;QCTyUkUGOW^ ]qGCW.UGԡUGՠHWV1Ǥ:`ᣡ+hUGIv +Bb(XMh p2BW@KAtaHW<>8"`r4t^6X9t z]HrיmOɹ|k9no ]ݴd}i߰}?~-~Y\]n.-S w4[(ϯ;-ovh> õr,4Ѻp4QhA/vo<˓5//V}Ig\oxL&}ybYXD}q~R~s:}'s?xf\?sրUMYg歓)5,V3 e:wfk _ЪO"Ql<`kgGQKXgG{:JO@Xh=U<]ucWDWWJkuq鹏{\祫ǡ}Չǡ{%+tԡךGDW?pEV+:t(w]2#GDW]- ]V:J]D BDWAٓw׸+euth՗o!{A~߻>[if{l7ONO{D9q}S7wE(Sg?K>>9xkԈ9pSJݔoѸx\]]V{bM%֐"f󏓛cs<ωB{Q9Ym^oW@],Oޗ\δRA7WUi7yBdk(:M4Q wIjvIϗ|0)&R*!y.1[bQ)Æ}L\עyU\O}l,ZWND+#DkB(ICdS6hF iۊ-`c~qhѵb!i鱆޽E֘sI:Xi<Ӹfz \5ŪT sea}cLf c3iX34fƦlZ+ 4eʪJh9Ũ/>=~<7yL4Y!t*ڣ[͑%caM:F0T`evN|:DRh*&|?FmvUEn"JD?=@#T'xk$rU[l֗̍т1OetAڳZfB Ǫ7_*KIo0L%mT ?*)snj,ߢ k v+**T(:]-;KkA9[ ja<*JJVҤV/%M23yAkt[׽`~,"da#(*S^.0F N#w060`ם^,jo_UAdz|϶u&h^4MLkT^TYd(yH]i--Um똂GO`#Y"N A-)BJ$RF Vd2ޒ$C` `c偾SH$=/*:Z+P6͐MB8ETgUS^ߨ<yǪmE4Yr&e5Dv ,} ͩm OO"{IYKOdf__b=b"A]jS6h!/fs 6R> ˇ$ @aZH Hޘ jwX1zV/ 6#"ZNҵcY aRH2g_XMϪ90F,m0Z["F̃`8!@'(Yn4Xxә$ fJ$ !?A.jD,8DDp,"#0 c9agTSp!DRf5Z,4]}t֞Ew֮i j@dc޺1BP)BMsVT鹿(6 2߬n$,׻E} f%/ݵ4Gzu+׸@5m?^sId T, B=Aqcc`3 5x (ؾGBʉf1ZEl(A#ZhAiʛӈ o=-P5n!/s?5H%Tep#`9#vI /ceLw᩠z넀0drH 3-P{2y^ V(͛+Ni*UJӄ8H&9˟uW| F,,SCaLrSSĈz02501xyT,`~ x-]ڨdr!cPUlBi=Wml6c&EJdXz=P*-AgѦw&Hb2*N0՜'еMuVϻ%4#%_qaAP  ɢWq d͈Ai ‡Y'X)A8 Z&[Zp\k+iq&0W߄545[zZG!PRₑ$5󛎷4HCiEjĔ-*5Z/J1`˒SwLʃf-$3.BD!6NSB,;S@Ǖ1[ ^\3y=ʏ0]G튅 "΃Rp┲{&ۭ^}#^tvvsP})%K[tdV?}R R^ǥXKJ--Su}fyC)~q<gUXCbeuNM&Pձdub= ,ċ"2W@#hdL 42F&@#hdL 42F&@#hdL 42F&@#hdL 42F&@#hdL 42F&@#hdL 42L *9b-Ex2 D{dL t52F&@#hdL 42F&@#hdL 42F&@#hdL 42F&@#hdL 42F&@#hdL 42F&@#hd}@;7n$vqјov${`FKݲҬ4ķ~Uxlq3r?dwZ&4f6)((|tBDa#$h@J"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$h@**#BgO zZ~$>m@I$#$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H"$H" tsj=N/+jJn7 7?tϚݏf݃@v*&FK=o6K=%ŞvKDi. !p郬>RHW ώkHM,]WD ] GWhPlI>qF[2Yu:ON-9 FhEyLSXNF;}K!wf=R0;Qv ~7bwjwsƶ=td\}>bl>Vz{k8Ya?/WE{Gf +(a\^TWb}z86{~o1_ghsr^Jt^cNfvL,7nVbfmq~Tf?d_ \+÷Ϟ:MiYUS7h*hEgݫĚ-߮`9ja| _׿*Y HJ=JX|a8 X7nx ӄ^[83 MW.$Qמ'ǜI#!Z;*qw.]q.CLEi8''fƞϤ42ij=?/^B<+6"ZSQ:JL)cSҳSd~aʢ5܏( [ ]i]^+']!IךEWO )!Xue*agpLltVy."ZSyueuZW ڱ6"ZJUKDWԕhHW7.BZהF>)+ B,`),dOghx]ei4Gѕy`+gp+t]A*UqI)wĄ3k5tSUr'eO:-Q'E'Q 2Sgn =ZKo7ECl7drzEOtE iC!+w^UJ^Ad| 79.BZ!J FWĬwg*_k 8*+<\y96422Е]5FB`m5]."ڨJו?jAHti]n\tAuE6+B*6B\]uEEWh,']uWt W:Sj # })]nb35(]|(] RW)nBvbY­ӵOOvоH}d%fSw6aV6BEV΂Ps,1΋w*^>+hIcӑ#ܾͣMRq!vpI$M? (jJTFB` |nd3g i.~'QZ Nz8S&d{Upt}”wg06@GbY ]Y.U86B\hZ4.j2g+p^M!֛[WDu5D]Y+Ogq\tem&+49,0Gep^A<ߺ"JEW8l֙Q;FJQ QWE#]! Tm0(] QW *x?GMlhw=Y5t Й_,pb !Jޟ%@QiTuP>j7^ vQJa 2EWDei0EW c~9]0X6"\\tEx]eg芻YϔwU0GXӺkʣ=(ʢ0] ]y]^3ҕI$o\tEx]ei EW+HWk~)]ltVuEƈ+4 Ǯp]pmBu.zuB2[`+u\"J/! QSg#v h*]WDj9ǩ3HO q:h-]WH RW)KK;aDgTT-t ZN'93W>Y}bd۶de,$\aY.1.ȚΐYFe_kLg];*0ECpquSyzqOc~yu1{s'z6i/*=r??uӦVcH3zqL3tn t޿|ucc~w 5ܬ G[Gjb9~F'?C ;+r\_Sy" Q{ aʮߝOh/L{;`j/t=^fYדOgU㛺5FځATM5)DJ5vZGc7*Y3=B.ޢ^ez[W`6yt=;sgje2'Ogt4h5U[9;E)!N׍6I:L'=P|ݟ쵪:eM_  v|ȶCsW~rՁzY{n|v?VoK$)xw 1cln:U.'h4Gjt~lv:K_Su.Pֹd{(Md\U1w3Psg4fk]Sx^s6ͥPzl>ߘHz 7 <&kDSaEhyX\n}l ͪٵ.wY/eUwe=뚔雘|7UR63+־‡K fJIa|RG(M'U`Fhu)44-AMjUv'kVIBl]['/ExVi0X]57wk?=jV&rb I65 3ͼ\^hw<;a5_o6`7 U5n 64y{wnh7Z:5b$F;kmݏ@dCvM&m ؾv2iԉ LdN¿tk5lԯkM}9Gh@K v)CHb"k/]N[EaxY'oϟvAp)씓~k_oY77:,׳jfѯy9zW-O_ozbl6>nY>Zk4_tIEIsRЉzI~|' ZvS&g {.٫V8mR6oI&)"Uw;7 O4<-UL} /LG|Z~)W}<rJ׻ § [E76~§ʭ(ஞfCm=ENߥjdrBi ]_=,j'D7լ_.LJ ķo./+-SzB#ɻVGS zx~~w18,!oؠ?B)nсn?rb6~ǵm[@Y;~_d|A|kmH a܇CwW 0w]lpٽ/{ ꗥ3ErIJrCH""eSS_UWWYSBq cֻ`HDqٻE`?!:﵃'׃j(Rϝ(8RԀ/融Zі4s{T99c:z18/qC6ZIqy 9xo 3}U+{T4Ⱦs_7i'V&l~#fO͓|s}Cvf9]UB=^6i~ژBV}GxYPoH!5* pW?yOc :L/nOZ~w)̧@z&NG=J^RMDMg#rfeWG=zye=Yu3O'L{uF̴ֺ>eP| ҐkzXVl)1e3J`t!YJr֜t&w]Zܹ)s>)G{UqYDQ0 irXbX@QN&&C]Ԗ0M` 989m `)D%<NHR&xYI9 V^+m89ùCϞu&7K^IaX?Q1/ >dmŘ qHTBp@޹$b7e$&D$5ɝ$ T1YDLE"L~h+: 9mڸO:-~;Ͳ$j["Px^sFN@? Cc[SNq8j5B66Ce9N 1٨ܤWfE,-? YuI2w#"D[IH^\597ŋ#Vy' fr`]ȯ4Ġcx' 0V\5W:e逯1z;)5`_ODCzH Pj QN ]ɉMKZyp j&yK))DJy"PavPq!{͔`D))p))3# ( eS%_RbP6Eg,X ycF"d#d0[ũeQqV*"xh]9%LP0DrL3-g@.1E[Ʃ>1P^x9qc`|sTQ"fEt ^ʽ媫s2F:ךsZkNI'T%CdH2A$p3Fv $ddl.j'okfk`P_OzZ.vZFZ|͛y璻>tϵZ]olGM#bS>c.1_AoJ LJ=:}mZ7{lzw߃OB \n 4%ݖ -]+L$ǓՎ|u]0lT(zU{*OLҏ˗/u4%os6xk mn6cx[GZ7}W*"?wx٣X)l=*;Ep:N拆c!xhR-cpXi8o}]NS^FcU:%AhTd.HA&13bia_عAx\Jbq猶ppo‰ߴǥ @1VB.z˔SJ  >^I8;9m. f`vŁYh;D. ,\U2]'}e/W︽Ysy8&3UJ@H.s& 2d$ז:M^@B1 hLȐG9J㣥'F霠*# h:ui#qy_;L]N,@I)h)I,Ah>gRpNN`Rp@"hμɢGϯRŐg#1f#Pfry)*rԐNT21@Z+4=# @+Mɇe ?V}&"Z57xU<+Ų1T3Jd4-K9" nAA4iG ӫة<_!PPY# 1YI-IPpG i-'rˑgIoNy \/|!y8.N9޼Kq-RB.P)mֺqðºĕu֤,$º -6 PsD] e_@dKx %^d>3JHIyYdGT'Hi}JQ\KCZWiX;}9!3O4XULHWrUr7NwW$?2F%Chqw}]wl^oz/ ËAoxB˳ š/u2&!=Ktd!W.| ֿÛ I4珋 8~>O5{ѧ-=YuEZzRrSVj~xW8Yrw4 <+ٌ:SinwtV;Lo{?o6kbޥ4,:i"kv#pR8gbJ҆sBͧ[b4^]?.jV1"ӵq7r÷fkRߜ0x۟dōDYnk$>0g!gr5j iqKOwdWv^?5~MZG6FB$tuSK lʯ[]-u#u#!OueS^vIa,1Nc8Q}g!TErlfr?ClLuK6uioY #8I[(jjli`̜csmn0.]K6,P3y^ԕ~f2wfb'36I=EqxFsga=ŋy@˒_h#n,nxth[} X ks^y0 K/ae0uWb0'g3侼6Z k̕ѷv弳ArZ,B d 8'**]r.0c&ϝkE c5Iсo`w[f;W;=9t3@9uu[ g9JxRwVDg-7\~VAQMYCOMG3.Wh])TTBJe0ic)n;;b~۪u*Z6po띙xjy D هl1̒ ,eRl2^|+opЦ]_m&Q.Y-D~ ՑXrIxeζ3',"$X0^߁XrS: !sF"@FOƳei Hנ<h8$I\>w>lVwr a:ѝPrd! 0ɣ)pKim%,@k_+TW.h Ô1Ȟae0Ϟ8NG>Cs' ́WwX92JxoVspUITJX9UU6Q'W|KiGkdoy 57?u̟zoSAA*Rc&WttUV(U?0ǣfjM8l "3Nri)L7! .א8D0w8ZqC4r HS2$QV\)6+,PƻzMLF9 k$ 6E^,7s N,/Z74.{IjכM~\z0EͯFgZ{H_!)Av=c &A6"JԒǚ`{n5%"En=Yf7ON[uojz1[1`"*jL7{.2>Qo]ɯ騙tޠ1{ tFZH5%+U>P-2a_6dO,ܤ|2EFgȚ$@&͒)T- :)fa۔.} {eet*E{pqQ 2eunZzz{"wl Y-Ɇ:o_6G7\g6IV; 8ꧠ.]۶3@rqIΗ]a-&6HZ}zxm$/va:CdPXݽUa я/H*olz*luWCL/ZwDʼn2u%EQUiWu+>;>Qx\PD3If9Di @d`}d-e̫3*֊v 섍ܔxgCu`bi M1M~QHBm"=*PLI`rTLڅbA VEdqcf,,XF B_y!g ~k@X캰UsvjV,)};<_o;fr! rW;yulpIe"I*oj7?XJsEl$gOgj-% Xgz( ͵ގ{|}ű(|!ҢfbY6mabR ?SLcimvjnPpTHLQe/|> ZLS2AK]yO_|0{20,i\eS gDz ttpَ5~}PteK"^Hͭ}.6_wv@a ݔxeQudh/~ˆZ^-;lAѧ/G "#ﺦr~Dqt=]xy:j,?Nj t8_գЂ`/CWC)[tAWz}[L芗TW" {CWVku"聮 ]I&0=XUJ2:]\t J{DWx:]}m骠l7HWȴGtUCM_誠բtUP1+ř}r `mzCW3,DW֮CQ*(v&Z-nV+֢/%'jrU]|=}Z ܨ6%<50bxѿ13z7N[#V#G\z*́{[sM`-Xoh/4]b JMAWqˬn]D?, _E|Oq}4H?5kÑV'C7yz̤WVt`.\G0ή/B_tϻdP-+l+/4P?yWɻXHް*[84E+G'6?{^LςVt~>̟op3{DW폹OpO_誠tUP# tw2;6y$|GF!^ rE]GЕjߦT"0U[EnNW%ȁ ] MXUkzc]ZX骠\[>ЕAus!{CW}VA骠Z@WBWȔ]`ޣD_誠}}骠4|7HW>]zCWn֮-gtUPztPbNrz`CWuUB筫R [+z^z 9{,{-FG4O d6? 7${"RU=z+xV BT @5^#RkoC\{hY%8H`>eZв>Z܊$ie% wF%=ZMr&n6j˗G/=gq}ǿ%XQPzDb4S22x6G?tۏ+o.8!ydGkX1֛5Wfɳί!Z kop 0O/XB_ \BW]+B r.~tuD6Wr:[ziф}Gx\JLLxl [&u(mUt$y )DQ↱PWn9Sni%= Q딮sdln 0'@WFF< ϕ$+ʥd#YUN]2F77*?_HMﻃvhr?MߚVELߗp)YULpf:"j#x;8ZGXkoVx灛$Y^sBE&#>(*)rȚE¾?fSes _u !Kb2!Xۨ5u T py=GH1PPh΄&ZKf6H$} $.eet^hjVc;r_]I/R錌s! s,)L,?{׶ȑ\bdpwPU!$F+Coby!4NUe<ٝk]$B"sOAha3c1^#1>anlfVHSzL-x~>Lmmǐu hQJAnd,4D2R 2@>R;ƈaT~61{WߨGJ<D@#.DIow}"!=sզ:uY(/`TJwC4*ܽOO͝4着e7-GpNTG*Y76E5ɍK6J3 |hֱ)NɇS<&ϕcHnJʗ VCJ 6Y ЗI buRƗ ZRv~^IbHuQCBNFG6Tق1K(#m'X tf dGh]udV@N`Nڨ/ ~ E*UFG>rJ rSoP0JNwh A^AkaFh^v݊#:CI0iD(QTzsPyUo<XX 2B{k(|-b#n$l SK ݑixu dF|Vj;xXUAұrIiژȆR7`5hS5B-E9Q G6S S'FWkVَp5TM`ܝ&EȤh\}EБN6>❤Ӕ`AV5*(*-"w@iLb oDXqvLv6B@F;=5(!ȮdށE7BnTzC{- ȸCLAAv[A( `-ePBdBE@i'rը |u@x;bςG ӄb :TjMPI!0Ej>dT j7XCy(i3@fih!\=+AtbE,N)(J892rNj &YRH( :k846Օ/H!j{RQRAX}k{FAyǒk0 !1T(`5R;%C`3d5>! Db$ۃB}EՊX{ɻ Kz/7nbA\ EBrܠ1*Tei5 !pBK.}G< dexwW~e~{6x[ ީa}Q7u`mFL`-$,>&gAuP<>m:D6>+MT$];-6*d`fPDdn.I. Ao%CJ$JET2R,zGpKRM|xLPBb2$kuk+O<o 1tXTuaR% 9աk4=!VdxaQMZ uY QH $o}'kWمX#W`>> ]{ #Aˈ|uC.b:EjC6<.JX3nD4ePEx'@rFEK֬kdž@ f՜% C8˾'yMV]ϗeXn.^.r, J͂ ]H77kF֞5`ӿ{B!b1j1֚cfԌڡbwU!MZSDàe ̀|12 =pe6|]HmhJ7a#5>dNғ֞fJOQ!, )nXI2[1xxA4T 64fs˥ZpU9]CC,Z*cQ4k&^Z&T$bnBeN V3v !KkBE9:W뉯XD;j0BI>z# R DŽV/nr{qj޶n{|j_jI%:go+fS7F?8}ϓR:/aIee)/T{}?tHfŽ8˭6-V93 G/߮V-C:un<_?ޭ~͇=Wk٫WڄS,A4c3ޭ?ꚏ~x]8?8ian?|q1nz^lz?fh{ Ù73\VOp/ɜ@fs'\F^m@<(@6H@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 tN g(srA<8>C'pI@BuJ@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $NuuvNNu|@ ܨsNW tN%U@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $Nu%'xtNħl@ fb!Q(Nct'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qx@[>Z'/ՔGio]{o۾E]\?bQ9ǸƸd{KV+ĸ1.'.}:UO_8]i<0]藥Вzâ]%]zmy?#`l8b:]1JBWGHW8&=#bfCW 7:uj2 ]!]Yθ+\&st:"r'?"J5Q>%kԌbͅ,:]1Jb;6Έl~;>mLNW@IV*8x!:=f3'%e/-,%Hj.g$=+ xF|.yo}WȆ х̌ψn4s+M*:]1J#HWkNf6tkNDWV<]1OF9]V*4xCў ?kO/|k?F=QjuPtej^YzwM3+S 4b:]1υL8wɞ ]1hBW6C+tl0фrp*b!΅ e?I([|%\Vω{ki.ZzuZQ Z 19SZ2Ov,bBWGHW|qb~6tpl" YDU?~m75jۻI^[L=y`UG1\_ .6w:~VHC-eI9eSRZۈ<_./-볌J.-io&* 6dw1],fIVe 5F]O{"{$;:EkK*-+ Ώ"SB(gՃ;vu" ~Fṫus+FZʠVsRW ]1\b/eeP,=OCW6%=+u森Ks+F¡A)m͈0G]1\BWֻC+FH;;b΅kmH a܇.#! s 8{$6~\S$wf(E _̐I>EtYUuUuVcWJuRFq%dGX 9z9"(ZB]\%*)=Q\K]*?iS^*wCrFZ]x V9>ꌧS_%w]%[a8zA!_U|nHԒWstD5+ XcIsDURbs‚Ory=ݟ+9x0jl;"+rWzاZI'*9Yx:.F&/[ˊ`PM9Cx8}?et 3͘!޺{!<UBΠWW+"]ӴQ?|?۶%6ݮU*=H }"X/csQD&^Ÿc4șӑ"!D1FX߿i} &R5ef9WP˝eiOR ~#ݫwx0߽?zd:iJ>_ފ꾧TcPVmsü s,Uu^蒔@܀,ofWy3)Ff޿yd1?ҏCBBe,YJL&0"~ W.k^_q8 @9jW ZJӪ73_wI&[OSLƿU庅ŀbҴ$W*9㿤=h Pkҏw"l&ҳM#ۊ^q6;dqOׂ 2k=80R1Ոd+:ǗDWganyysئWÙgW e0;_!\c@8 DO"k/%%\BDpCL¤Ŝ(ambZ}frܧyzOM@& |9%X^-92eukէPAʝLgMm}UDuD.Ӄ7V jՋ[}Yc7ԑ<]M+5\—R_ Q t_]0-~U{f>8g;5?O'-^kN?@4v%$-^5a7Y`3fz e*DDo(*<-^uӗvAskzS2"ۗI$ >W+-th}| ~C3)yL64 MmlE V{ S-*ofE r^D!khҽ E=FEڀc}}>#}~4>gsM(WwUOA=&l0a|l2*A"ÿu D%7m}mcп%@`QE#@4yA]`@(A^G(^yˆ^FcTBXYL^jʈhA #! H8 Ӻge7rv+ָ 6nM4{0~>&c]a. x ڰ}kgtϨїzr>y(F S tdW[c<]`HEV*cdE$'$!r'p Y3G#(´;Fn6 WA4d?G$x)*ɻ͵UJl9>fM$ aCQ"vD H{V6LDOL`a R(I56"$x<='I0)#.2 kЇ Xb@ҚE$n$E#$NO>x񫞂㭵FH7J:DBXK/@&аDꀙ{[kiw,Тc? " PԷf7 i 'ԇO )t x3)_%K*5bo(96Ġ1^Xr<5KiUGPnq1>67U@Lj*JN0y00x[ΪEWg7vV$b6O, yó uiZ̍Vƍ!M\却&b\.@ex4B*8~YOVh8ttճ[ RH) >2S?F0nsCהS)O1qfՆڳN'{T/ RuޅzZK2Axj^wUTMoh-ȪJyP)} A#KǬ md&-n4ec1݁ڑmeM@[Ť0EjA1u3?~QFkNȪRj0ܦ+}DF~Q6rDolqmPFm(oUA*O: QB͸݆,\6q@P1hZJs_'/}\rM1J G7Λ+͊2[iBSeߛIjd Ê')uFOLB|tP(һkbbG<]O 0W4جq7$7g2?η PSj7{wTe~1mE&Hlih7 VU;ZZ~"vPM;ίnx1S":J- pT*1e AGM`GNMR+o'L/|3Y_cӛU q񪩺T-4u'TDŴ#(RHN]$ΔKQ *,AkOB8ŋbHשl7nўN".q9ząG/#Sb֛fvw Сr½":S=QRSjgJ!O䦜û2le_i* L2lP&pARLNӧuϛH1OyZO +"^9,Yhܔc’˨UҸ^] 4vY4a#A29HysQ 9*z#go; Ƀu"aL9km=[t~n̖^خss0Q2vl.0l.Է|<~cwT^+"7)&N;Z#QЁ@ya$4FoD.p8-Q1䄥ɲO;hªAFyR;-l=/9tT3UGXWX";H0P3RzXPY9%0n=!ie.ǹ`NDz2O8^r .$Q3-9˒ A'tzKR3Jy)L|6L}G$0#OIL-e6;Fu˦-&6 &$zʥ}Ŀ8 n \95ki:vD`;|vX$}D$]Wg^ËeɃ_Oz8[C@Y=Q;9[sέ[ R=q;|gܞWzO;S[XX+_/=:^2(:o;cƌL")#cҒپPzcf,[2mV?#tlS7|]T3D3SϋE-wzDat4e.#K7CX2Uƌpeh:wC|E/_!Y> a&\Φ50&੢U]V!M5T=}^tZ]>)ؚ0 E}u_qɵf$j/k3Ii`7 3// &{|(o5[h l1FMv[LuN8ٝBl*R#L=*ܑMQ7y|9+ gխ9;3H!r+iu.D]67IIܞdg"=f/Ehosln{5 igYsv~g--J<m𵔬?QqoܽHR z~Ś/ ʹo'z9Q%ʶ6! ۶5ysB[TD}pͽ%= &UW<23* 4HŒO^`d$:eKhi߄fQX*O3TX"7ZO)#>r-r "R`Ʉַ CdkúGp&[6ujd\0$YIЌ@=SBHX'm9r{CRV@.Rfo>@8`H)Ű[)r}ǴE.*`iڝ7>yًz7g6;.S* -h,*$p1G4Qx$X M>j$3bnU2 %ZM@ySDOp).}Ԗx93aߒ7rvKJo7x,=BqOa}e- w_i.n [-7[,iIg;O5D٢ĥ%6gHtU?OOU=T!ش;Y04(MC.zp;Tv̹;bdzӼF1%jwFΥ[j Q[A>iJaWcM+8o@48JP+6JA_A N5M1(yc;"9v3x8#-0JD6?v9"Ljx(EmNibRcPPC\(ĚY{Rc B8j>h3řVuS&m\K&٢M[%.NKms\1.qV$ 1(@/˿/S,=>&*Qκc\\Ńfǎp=G<[krϗUGyw6x?=Բ .'~gkZr!~1++$cIg~6Y僾Yf߳gX$ Zy? `BC^O<||a>հaY(ǮJ5 ˢu8q/g h/G~,o<{..'ea?p4/py^+'UYwG2uvKS'g%.Oq]63+T(RxLNvH<'XCӳ!ZzN#l韞Xl,y#ɲQI-'`9V//!x}~eCYcb5$JM Wey/c¶zv *Yio˜-qv|r|I1׮:]?|{eanym:_SױfނlČ赳|wzJ{Ѡ}8?2G9x\> , *ERЬ00Ҳ:Y8p )88 E3 dt.Av^ېMBIA-mUJ% mg,&gJpPMFwgEcV^ $sUv3g:I4O>|<{6=~.f˶~]_n|ϣyLwy۱L:t;]iN&eXJ֪N֧](3t6ȡ!-#XHx`#>u\ɚNN͜xN. qꞀ37XÔxS 79}?,=ui wPe7oqWN<"7}zvR#2"3`@ε^u'1V9&ECHC5rL G4G6N+B d!eYPlsLH">]9aiHiy#I6kW্Jw}u`f;>\;~((a[CbA=F@3PmNodY<<;( [iH(_x[4~PgRA+NN A UFTUR:REUO*Ckͭ#y0O<>O/g;,Z-s]WIyˋ7szݾ}|$tdj9e#T1i4ʹ[֞.Qbf+ -žno&[" 5bd` $F*\7c yT5ͺs߈7A^{5:) 6Y>$`ͅȸNNM6F9D;8Vq͠佒R֪Hw_tW֟2Y䵊2SNIgW@/A⎉U^g Gos*@?wٱ< =b؍tS[ȡ"9~ &Sݐ͋}XЧ\v0osoyJBա 3 qv@W@4o8L6^Ti6m&N}mNm;5ͦ,L6<唸 ٢#JB.)-Q_B ci;_bmXL!#GZ.:j'Rݱnl!S'1Uoj帖 4Ț&'=Mb[$ۡZRB-f_ -Ε*B.+PqN و>.8zn7]%?=9\"B*2dXT-,'L5WjMy_[ҷ76&L5Lu yOT@b)*&ϺpD={}D_×Zjjl=7bp\4Z_kI aцEnʾ' 5z2z8d5%h dmʙW- 6ei \1ۘ juZ䀦'=,~~;dBNbR:cV-y((?գҊ)oCL8iP݁k픍_Ys@f̠@qB>k E^1|`ѨP0Wf0aX[[_*evZ5^ZM@6P暲V5@'Ψy0K6w}i/sg'Z=E mvh}6[@,DF{W*0GbF2#CյfM-(4Rl]˄u6"%P|.(k9Tvze7/!))QRu S 1\eX[xșwwUE| V겾g9cOE1GS+'C& MKb e5duHp;lb ӏ| qCNhPd\YSSVYHpZ OFgeMJ>v@>6)i*{ھMw &Uܲ4>x{ q>_SAfr1|dz]kmotθd c'`s}"el:^cв):2[Dr1#)*J6ZۺWmN.U ꐋyeߨe͆imILNXfm]a!-ߏ^`ļ=9%r)aP ZezW{R}!Pϒ JG/s>.S֍k_[:4hֶ>o%qb]ߙ{,޼o8hgjᢌPmU"c5 &#%cIt>Mm30{hG @&}^#˔;e"@ R2sE b7e1StC7_2R1 rP4G(UlsE%Q*q~c!8{|m|)o\l `5Y[h {@US7Q%Ԭs|ϓfZ &Z4B+Ew*D:kGG(B8O+ƌus 0+bBmm|hYn2f64aJ@Px2\CXOxжu: )7C{+k XL%FGH(rVYqݧh*j1c0 DbSeonoanpq[2I/qzffe{NmA/Gcq vAnu(r^-c{7\hX]Q&Zۖu/'ϭ>m$˿"S7v]Aҽ3@ccC&ꌽQ$d.[ttPeӀX8į&D`4\;?KB]e[ch1 $ V7_ ВVTQmɧ2Owzm0yj)Ksf_eg'0>Q1WgӾ|Yaۛ023 JOrl| )#~J>W׈tӻalo 7w^ BMJt?[~dc&7qM?yN6d'a(TM!pWFe%gGmv\+ݽA ^QkΉa4 UdA$Xz)Hij%FթխXOiT.(^]W4! NWgq3t:ZcqmϦ䜬L>B)%gîJp͹ -']gW%=zJqI9 bgî2Uغ*CY{vؕooFe܇>qp #k\ /p*c`/8Ψ܏REq3C|]km_O#\,xQ(CWUʨ\f!x_WW]4ǟQc%p0?鏿yua;W$Y!8 RY VBGXd;P`W,Kewг #s ghiJޟe]uG\,<`W7Y ;B6e+ޤިBX g+HNίMί}$+CbC}!6$I{Dh)r5SfS|j[ 0B]2 BӸ^3o Qbݒ \kkvWf![j'Zh>kCTV!PsHTET(%2)XhlsvU{ڴstAZ  ݁'r޶mz~'pn}IE{iqه|[\9}#urJ=.!&ji.jƼ3&(/|tTȬ}ۉ65=M;] ¶״Ni16xA-i80MF249l(pH( R^:N(Y±4xX2Nqձcŵ}_NH]9PHT슫q{Fh랹sgvns\(Ku6&P)TdcSF nunA Up[mfNrq :z 籄4x(5fBۛ0'2a;xid2LѠ4^‚-xVA2D㙢sڽP3E7xU ##$>hYTs$$r@&Doe-i8Iy@Gñ=ޢ)DŽD3LKr%DwY%$IQ8"0@9>K}7i$gZE4N"9&h2SJU4~X3z66OCf*QZXK5@e*%[1 CRj#$v-WB3ptT"m倀 t:GY2 ]BjJYx07H.=*K:c (0jgpQcYB-^ܔs2(,xkwhN ?} U\&f1`+d)Ni *q\ͪnwMUZ٫u;;A'{ۼ3YkD.a,}IsT/ dpXRRç0~[:| [dߗ3ÿwx㢤Z, )lAr%o~eQ)GG|Ĝ7Yuվ2'A=~P7=ս*pՏ M(R ۛ\1~[{ߛ\dj4V~.rӨmwkꯟ'Wo7\fqUjJbv9R d&pu?,Y`nD_s.` `Ɠ/=>9´Y\QoH.W}G>wCoXr Mx.{H[n乊W"yCdr . u~wM )glk;+<ޒn!%;K:ѣ8GےOeZR`2 usR&,|O~aj}Lb }7g7ae\g@^J2SFZ|(^who 7w^ BMJtqݏ,y$SWٔS5$na/=hl>ٹ"O JGKQ &B2ܕ+se2se\T@YZʐ$D&Jl0.ya RRT3$7g:7XXuh3$%X6Ц0mb"0<."I8D;ѼB胲ZB&r ACPNX9۞8߷>3V(_B\q"<+WʗN$6L!9)E[fr'¸3&2!)gy@Ip=F'C="xM_T(\I\FTI 7^\ |J%eU:$Yg*')҄% gI r("'n;߃5p%cDůK)d.2gV8-Ht!Yǎl5%I̡% <7N9G*#M:$& |~*_3-9g{T>+ax(E#I $$yYjjQb̦Nu;=Rgn PW8Q)JFPbB#jhq8:o 'HlYXՀ`0FdΠ>*U1Ay.I@`:PJ8oz_f-gzk EF|XqhQ\>jwm$d'Q]N8Fgu^l uQ$4;T_(R&)8fU_}Թⵗ5@21zE'!(`KM{gr_8H@,W%cy<9Ol@7hYGP+St|y酞y5AR-(%N>p"BjPF0_ZDTSy2]~Jp\0rUQ 9Jga1 S flǔa[gHl}X7EϚ=  Ifw ?AB|6aMlsƷ_,ͯ֩9iEQFx) i*(X z>vt۲oDiߦVzMw  l᷃~B pm/'8ĨүjԊx_wP4t'%˻'85὏t>}mgr3~gٷ$یj򷭇cQSHiy•]6 τ~]gJא6 @+u'E-ĿÑYxPV{5 $( X8#4D%UJ {8w9 ޏ/(4[y0?6PR10zgS![|S=`*M޿1h?%\tY#\Lq= y2?y=]ʑI)".q]ԒIQ ^g5gjϾ77k: i#!n5$m+mx߶٘6ux!T56tQvz7jwĮ]{ht:jpY'.n2?Ћ z9q3a-gc؉s@".ϻx߿g|z /]3_d T.g6p"SBXD*GZ;M'rgA{Q#v"d)mu Z! ;Gw98URGDz[VO@uH]5D"~@mhiJ#5~)q(9$=ey;T̵e/, 9wބ,*u0<Pne{4I̠^z- r|Qyzt\v+&()|NT>ٔ,'1{fI=  ώ%ѩ YY-hy-HU`^S0:1/?)7QPx"dV5IDs< Bq) bmBRm80$ wD"6S8ùͧ|>Bu|]]yܬjNWv8 ƃڢE!2an}u=!od1qT1WhzFe/$ר?O|S'|Xؑ 2_Zu}p2A5(MTJR4bxY@7E%,@ӚzxޕO G `$7g y0Mdm}J)Ra[ U3JE 79ԜpdUp֛{1\ܬrz#~1{a{Wbc϶?DԷCt7|2DL;L`NHe4К l̚ hWl ZKھfCHM(l )-Ct++LW *vBZtutKVt3{+D+[OW_{tj9)}0w_7~x$@CUqctvfvɚ396?ߪ[BOįrB*Ag^F*KvhM~PyyH;ӫBɻXt!E.P"-Rq> ﭸ(&gZ4TM\1+9|H 뗃tQ=wdA> z(Pb%H}VB!?Cc_vUDvn@v佛svQp!B]!\ә 2t(mD>Jr;DWt' p%]+Dd QrCW]/N _NWCO7فfh= ]5Cٶ9/рDOWzOFe [ڮt(5i͘]`NLg t({(ž5Xϡ++HW Ѫ[W%Jҕ0.YWX5+th:{t()魫c+iCt ]p; "ZNWR1u+thu QZOdl>zH/kIxLd-h#w6]ST6R %X.ǹA|^~5cV}'?m'_SMrs_gUa]r L..#avRj~ m)S];`!GΤx"'! ~V(9cCtaۨ}3I)RUTݱ.X]+D{~P֦z:]{v++YW *vBtutʤ;e]!`K:CWBt(9J0mubWMg >klQthk;]!JM{:B%BJt]+DkZBm;𠧫g+řղCt8'Tw.uho;]!J){:BlNCm?Q4]9YoꞹXEk=}}wNhɬ=X-?I}v謨#S̖\}g T.ˈkd#ѭ w95ivzϢ2rQ:Kb߿[8WX6>E i706 *<^mIkyMҷOh'ukuJL .eL{hLoRCqz֐yU{3xH>Q\-'Bd q9^uڇ+ٯ+ Ju{Ӧ0ipIx!1b$DM443\KC$!hdH%mL7 G@/tW(oK3<1}>8!(^1DaRkRV# '"xeR11i5eG X%Q\SP`Tߊ&&tl-xҼd;d-?8^]vfr2qj{aҟ*̟E+~.`x[oʹ9"Q(`HZb}LYM}(o}nv:k%% :"? ٘sii#)iɩ<^uPrqư 4Ef*1 zцitֆ 'Bk44"9PD Bj`+5R@p$U%ztk)_{b< n|>/nȇg7w_?Og~ݼa0ϗiU:MtCk1je21>- Nv;t=˫[ez0mkN^e!]@]F9 GCi;}GduOY#D?H*x]G"w 8WһnuJ'7"m YN·@\[ -ջ % V:Oؖ>GE0 &4.ȵAÝ*Ke>{|cz﹃'IEE=ice5ve'T:O(ஞ稜Ɂi"'S*cQʠ=B)=ɄFh?G9yZ,/)s|6M&MȽ&8j3SUi幛&7]"W'σb|,e]px :_,.e?thP0Ij\dmZ/ U}$vX4Ϋ6f<)xk:Y ʌR}-n)ĵ%C +FW,@em~a-{ϚDSS$!BgpQ -q֕Gm!y'>۶ Iړ! ,G!I,2t\fFs+&*ay"\X'R ~V^po4$ΓxKIXJ 4@o4}{}9ENe`-cg2.G*O.F,.f@R$)Y㸗^¸61!K+xI{ݛ殏4׳S'b&Wgi f.[Op-D 9$TV$gdu% ȉ=S,eYU,K#ɘbeglopgdxTݖiL&9skA1v՜i :1G*nN S>("{#g)12CN(" Q& \H6R@b6 5MZa1no; ؊hP9 eOVPC#j8BߛL sM|a[k)LvIG`HU4%:#``68  7},sYai򹀎E&"mSxqQN!j!E`H}61f぀I,c1Ɗ0jvMBfa1ԏ&w#!_4$0Kn(7Ny-; r3$JK,h`୍c!TG`? Z|OvLt#>\ܿ^J0Z{;LF9G-ⷃ8ɱMZfm68ߎ:n32mg *Vfؾ9p!=A|v~Wp!,FНKVUp>+ߍ6aqYv>{ ߚu[ju=z! uu=KpUxT &WRHx+@WR%+JԴWǹJX])ַRǫwLǿ^T&Q)ML $N90jS2C j"QO]~4#Ͳ)aLpn{ċ?Jįq.le~8 L (T @<q^]AEo3o$Oox,vqfMWa^HJ1s4c4mC%<0k'ATYgL%*Fs LгG=֠~RLZQ,P!]q Ab32hRX˔c H2+'zx~jl?W2i]*0<{t28=:Z[V7 #H| Qs*E 72G/ 1<ϙA dTF-xB)["DCAȲTUQhH$Ҙ=Oaڨ59S10g + LoE cmoF5lX &7Ԙf>&+m^c2#s4Or}-J"ᥱ?aSUAyՖ3g%F!P% 3q 4FdHMp0E\_^';ԇ)AeuT^ߎ@gȏ3D! tXsA^қ"pGynsoۻiqӪ1 Z$)rrbs@]RdD$4XF5RH#LT)#xEEO%&y' $8Qg'(T|E}80?j9,x%Oɓ {8{aw=G'zFh 1im*r-;ϴ*/O2lzC=/(2Eԓ@Qi&R !7phs8|g.Ɠ.řڃ$&s`(B`TPm,1~XB봞Wi:;7.ӻs> 3}hp&.$Sy!.\do /\]䞴ŵWgHeb5cH9~OՌl , CUJޯaD-(qO3d֞JS߸a7̞=ۣ{M7vψ0! qUd\\fwU##p7vɬUl%J)_P)éf|8Kz.w>uxvNב{r[0D^yN0qd[p{dkw=o0'D[D?&m'?;5w@7du,_.FD?o\oybٔ+Hi5k7*Ȗy#Cy4`d6h`ᄺ8]?&|K_qa}.{<'R]J'dAVqB3*(uݴ ;v^<-3CwqdZsIw E2Zof;\os.`2ifmxd2?\dnƆyam{a3mBo?Xwd)khYQ3%lPjL/s/жw7J^xA\2FA\ lց$wjk36"ى gY/@?\6}}hJ#)!$&dgd5`)E* O-XM+"2t^`~=9xN%?' .fzKUˆVݏݠ~_AZ<Lz:5~`4멫/S}Xpi;{;@n܍Mt t?×>_Z70=zP5$Q fMWRV b܄HFAwc@h;l ;b J[Rt(f0HqFmxf QCX" ^:eUK*uΒąN E/RuOJ6C.Ed@*,hɄH)KxJl`s AyMbvrCR$Jh1A=7H1rAY* $ 7w' 7 'əeB(>!Ghoclg]i9*c85aF!!肢Yq+yW9YX ^N5)/xRAUYkKܪ,0Pl1/t h5%VFWnO9)PB t2,QxF-)b*1`$VVKQe$'fX%1!Y8(Z` *|=50q=fMGlp5zap~^h0v\EFoqk9><߻Y>ƷxiϞ?3o]@0mP+YWѵ[RRR5E!USOq:ӴZkD>G}AU%2ʣ =XB׋Stnm|"!!"\2Ap 6mD,S'mNg"u+lP6pJ&fZ>罥Belщ ;xJ{Vڨ].!.Kǒ&h*IlD)BЛkov!ѿKFrƓkp<\xPS\yvO]_D*$p*TSQy*o$ YTlGm`ƏgaNKSe~pwl`.fߦTCn ;v53Npӫ'5i@.3k@|P%rAEMĿӑQxP(yD6Ed "i2NC(w=_Q3l]kpdGUeOj}N unwx?[{t#t cyw^udGڳ9;å&qWJ} 7m~_UUӺ`H~Qޝ({ܮvmz|߸DZE+-t<y]5ߋ72˫2-K\ tk}s f5; <(Rj-m/diFydK th.f$zvUG"jVH@ʙ$+?p1{Xe$9ʑ@Z<I@(OD(AGX!$%aZQKto{n!Oȫe: /#=gZ)5^s`QQYC$ȥ,(d[I/[і}.TWT7< q&W:(ᕈ٨vĄaU 9FG9%⊊y*Rw* EZ_ UJ;S.Iq(@c&p I0IFIg-"C*P4ʚ$(܊ e|L2h h b얇RH5HbqD4%%b/ǔwZA*)VAtzM$XSt('t({-a %@~&)ZL80?{ǭd௛NXdQ6 p ݽ`YɎ{83z9KiIV;LS"ˑ75HZ7g{Gso+:.2is4+ٖed^g^ ,δ&mU)^n2JѹMpeHMTF.<F'e[1Ъ(lI7g?>Sc0zҕ]9<&n969ԋw#;ZGL_L7!*u1,|ñ߷|ǭz2~c 4c]qXRP/%Zs;&Ur7B;qtZ *>DR!%R:Q‘- .2BWU-xc -׼Oݶ\<zٮ8~[uJ)XRҧVyY0rY>^Qf?Y%t*x~,v!{rY5knXf!8BywBv+Dq Jus tG2R-d*:5>D/:m Y:2!lfRcRCpLP)YȎw)tq}ڥCuw8̫knd^ߟ025ګ>J=cҊeeJQsQj^g}[Z}y]BSM7ofQߺ9v}U||G}wxz'/ R>:oǨ]m,n|5GT -._[*|W;l G*skV5ʙ ךv΂wtT 阜)T5[x16f#a|"5n`eʔHB/i2ml57,dؤ/6|8eܶ/6v~;0nli4-xp)nwy&h~4wpvrT/YFCsoHBiv!$CӶ1kQ"haMXШMB٠X{+fcs!qPIRTr3%(rֈoZX6h}6/oIO4"YV/N_n\ߗEn^~z$F9+g4^4.R#}?7J+zBbQsd!qI2ݣdb5æ%T#YSiiKRj'*oBY$Ŷyo?0pinP l5lzi;_>>6UM1KgL6 ;:kO@P/Z=Io/o?4t!>dn{t`baQw+ ן }\ ^W&b)d)fr-} 8:[*WW7k o{UpВp|jnb ׮֗pQ]~Y}/2x@]\=XE㓃~`=a{VuzѾEx"ָfwtq(^}o\>_PŻ1z@7;k ^!(WLSf9Zh9+O&s2+H6]TV}4xjNT* K2Xw,+ ÏkmC7 2oW+oz`OmWR-&i>zsVS&)Rs̛% L83 RHQ\"'1T9(-UJ-dZ`mv&bR>t48w[uW0{жad׼Cn)V8m+慰1=#K4_{`3/= qϻ0V=BP:_ a a~^v艭F-,t2e%1-:jͱqׄeC.OH^'-/-p|-[ڀ(BS~ZƜq=;<=S/:wd6^SK1 52~H) Rg{. ]k@7nf~ۚ[e?\$Zfb=h>P597?9Xn-#;v (dZbeѺoon;/j:;\=E3)9!y2tN}G UGUHWț)kwn:LvG¾UG)n=>|l/y`[]= y|퇡g r *tSpxBtՏTxc3hY:kYZ3]!k?a2t ]-+}҅^ ]!̄L0Ѻۙ^ ]Ļ{L0þUGoK3]= ]YqhBt!dʊWFMޯ]~HWN$),f:ÝhUGdHW, ?Q0V㓳|G#݀_\]?ob ./wܣn,(2#kc/M Ƙ4pKA+J~'2><;/su9p8=j;YlvPlbxxK9%PUhw0^µ=eȼ&7y(yG6ݿA)%MuOp*X\h*\GGrɽHNu~J&CW-}rNIJ9RA 3Ճ:9C\l\@W2Cwj ]d+ۏ%O W8{>ޭ߳{GG`H.eS뤳 ?u畣I(U#釫 zYEh8;?Y\/˻/vyѬSx`ߘқg.'+}{ w)vYe쒸V9a#uQlUgv5aU7_R7`^HP~cnz-ekH>>Coܿ]6߭~ZsYx8<.\NHY)p -yfDkt8N.Y&+Z1+:}mz~(ɋN11jT ]H1bQWOǔfpc!n^jFZRwvAqP1BR .5cS|"˵d92:DoDHnLX+,TR `L*!\IE|# Ŗ(69ۈ]r>y1$&[j\67碛L$()[JRMse`{]Ʀ#X34f Ƥ[+ 3gUbTg;@â ὾z_ˆ&S0h#I+EGh6o|OWP5P:jJa0 3>Mw!\1ИUZ-KC ïUGċ 4bE oI]=?=l >L:[ȃ%XRR.Ē]5 >>9?on@}VeNcmZϩ W#jFj΁PZQ^ܷ$ɇY6nW)V1T99ǯ#Zk|1 - 9zcU.e\HVŰdXdN1Ruߊ]*"*,92lr6U1PFVM|mE2) 3HX6U.Q{5 =VB|.%Q0Le*z˄P4>CZ6Xv`1O- OY`BQxi0fƗYqYAȩ [.һ Š˓±D6VSF\Nb".%l1eCƺx֭?SI9pulj_Ni\uTx*P:rK 2 8s( [lne*URXB0 e ldd1G!"(N%WEIP?T(@vUG8``UJc sz%* +Ԝ eq2Fl6Ҙ` e2m RiLqs6k)χN[S ,,xč:M;*H10g_&ses?H!P:8%@ `-~׃5x&# pv=X"2TtgJAQŁ.06ETS ئ1Ƹ̮v9iqR<^^炛Ӯx=<[ݕ>t>H6]`-x: @uKy@xc#hJ)hGHW]HV*ea`!8" Øɗ9(XvAzjN$rA1ː>xmBV5 tq?XGiH^Bd\fxq4ХBU+ڄaDPGaCt)}=‹9HbD :tIey B 05N0AG ܖSBFkhg];VƁ<DkaP(k(h 4f Ee;]69 Ф ȚL\bQ0EDiv%& A*8 wGzgUEתwEAðDH3A6 ;FtC>(jSxKZl@fHڲhF1MYJD,RPt[ #`ZFnCHX@id_ϻh1-i0- ڪ mry?X~}9]tYNw<2r :L%m`0֣;nJ''f1za6R f5{O(w%[ u@Mf ScѪȴAh` ` aвM}̀KHIѺO3bJ7 F{s|d0({*H=uG!G`Ie(64qEOU懆װ< ߺl) ʝKC$p0r(`tƒf5I"d$~B=(jI@( m'3a!HL ΃rt qP R/*f-yX CGkv!)95]њMNmKoe.s.m W'*ϟO6zT[!P`7Zrud2Zi4QEo-Z,~ zs z16_MWomB7zm1}XpWrrz*&DͧtL%i5[uvov~"NvK7-"%#£WRzJ@@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X t@QE/I h{0%jq4J +9V=A%@J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%ГUI)8x4J h"A*W/F'?ΩԔ/oKku9ݏjJc.#\\<%ZOr^ѱp)zMN#+XІPtJ;""xIh|tE(`zteᘒAz4tEp>"GWrtʚ*" ]\kWeLWO΋#+x+5>$V1]=AҪtFEe׼&axy5PjncbCťg'Lz.e9 5.X<?~emR lf^ȿp͗Ӌ Z/^7zym P{yHw~櫳krS8 ]4xCj<t"l&(WND:>7T]5kYQV.BW4icj嵼Ene0oџdDpgP:Vdbc۹ZkWXbGŅkS6;=)3K̆'?t+3vbDѿ ~j5gwXʱ#'Oz46Q/?V#bix&hʷ{NJo KR#+vx!BWNWr7p>cO3}JWC癱'JʷpezHKNӌttE\Y u<ś4縣[/^VKgT%j8`*]H{}WN3\jͫZ;bzŭzzAN@"ylm楿|V *&S֥x'v"i#NKa-R肘Ƭ؋Pu a}q2=#UN֢4Á[;@؝ 7p[ ͍NڗkrsV Ыi݈:Xؤ!56F,m "+o_6>̯f{ !&}?\^;bQ{a̗oi}3ǝlV>ĈϏe]r ƪٺ^|yh8ecgo{t8]IWS+J';Z6+ljGDJ,-o@g{lnނy e}.^$qΖMwOd oAZ\D\7]n|fK^W9`~]o9"}ގɴ1U)o:q{o=}[_{{-_ozy1Fmԇe݉TR![i H}}4G 6oC:?&g Y?e%t-D&%S RDF$ui+GK%6bT>ȑBCӮmA N AXɭtsQIMu1v:R<]{uVA6$R֞).Ht0UQ*% ! 홦OyÑZemfغˍጦ%i2 o<57dj5_#ti r6jl5fQ7\rvuy.A6'o@u܇,Oo4tz޷ν! G˒y_ [Jވ9% +505sjHWWyUÙXV=԰?<՗uUER-MGɥ.g )[;ajzbhwmocKMӋmU=O&ݷ^Hm7?|E77~ <{X`G^ }?zqyK/F]mUub1:^N}`wQIsR:sQoJa?ݺǺq/xqa{˴\7ǮImro>O1v&eȳn:-9_c[G'7&},tZoqn5.5[aalGe+Ͷ)u+?-棟&zFn;&,溽eЀ;e*0o;$|u,h&yz;n1W ~ '}oǰ~5ĺuw]b~;y% ^?rb5(;/_3߻'}9_u%a(#mH]X\d.$Ur R܅. K'Oǽ1Nא(2R1~xU6;Gϧ\J+?ܢd)YD4oH'-&U-> {.9 YsFu5e[]]|&&gMv//fm$_~uYh|7)[xgX;Sto1lmOJ#*&My0ja<*˃ ˃PyjS*$)XH\81:!\!N|UHx*$ITHʤ:UT)!RѪ:udXS}Cf8͝hqH/LL"~T\Ns?TF4(kT9-RbHtCsM#R+xeCf=^o`dj1-6; M !1 :gYJ3u.fA, :WZip4&Y9/TgTTЮ`D*8)|=E˲\X)U2rBVNhi)Ŏj4t.pHxG5Q?{֑ a^r_ E2,6$3/1jkԐe}/"eIӧO_r/BI 1R~N&ɘ(J$\(ARW4;td)$CVtPԊ91NO:-~;%h["Oмuq71΀44&5| *r"}D0&:x£:q *J͒):w@I:yl-"4û6zr.WD΍H=4"hXtѰ%AL UZG!u22czcM< wTCU5Ao'`0推I6iMH/CBw"8gWz Zk^lKJv"IxFRh {p1͔`D.)]R%gKt.)[PH"nמyhK #21\vFBUY gX "֗c,'$ii9s‹4`2N6q"f4LJ3M d9•[Cnf5@97DVARl*TKp18|8%GN0YVeEDR蔀яQ>/4if"2]X5NX#" ;Q LubL?(.ey`#W\k!hJzbJ E%"h@p3FSΤXcgcc;?t Okǿc7efcPg<9;R饖9p)CZ~|,Uqv@̏K3Wa{]7o.F)Jg07ju=o)a@sGBrydxO$ST1;&ʬ`J]n]mn_p|R+]}wޔڕ#N^omZ4On߶pt?r [ƾ)n+?•ryJenуE˃^L0q:X1测1.f=RChӛ`}U2|@6W-km pжPka! -JpRuu};Kp5vkh .UWZiF M5+-q;:Ou*uVjyxqر?u9LJQE%Ya܅T"R FGpKD2/۬K2h S.m2*ͧÉW׿FǥCȄ"Ph, Б e)2%!LZhR\AǛ%gNgNA7ZaA;nq /3' ,9SQGd4haf$hPT!L;e 'P 7'&D#Id>Ii7כ{֛8~ q:?:_*՜ ٦3;MZ>l/<]>⠏is9h| ->c1Nn ܥc:xU_Dzq~ehmxG4ʹ z\F*E?lF7.[6s")=))kSvT IT`3$Ms}Sx4 }BN݋=|cʛvOa[-]YK+4pzWoȚa쒤ӯ JhT~G=S;ud;qr[iӒ w{>#f-5?,3|9,8~g7!k{rerô)~yHJ[y_^m6o#!zIs`NE Q8qY0yL:MsKñsX #aح:\kvwN͆g\os.fܿfmvC~)e歞'ojXVmz3jcypM.:Z(]v1V2v.r]d5(ery X.vwN0kA ;gA*Yͭw }zrf8ǩt)Ub+}\U[>c!f (2C%Ha9o h,KeʐV7 3!J턶"7c% G RrO&t]<'ϒ gO}k)M8M뢌OPah'qܦ~-/|wulzĵˣe{>$S17ʣBa8=:q=C`V2us~8TEd_Jl/ccLVykBt.E?L񴽐+dBO zhB'Up h!XVS *?!m:[: SJpDH(ON%N˹ 1qGJd)I Qx& ޣS݃w)v)&A D8ř4Yd=8c%"38 8K8mxm(2D HGH: f$*b(ւJ)E=3q+j<18wڋxɡGm؝*Y%* F7pk-$OIGHH-Z QXr}NQjBH L1נB2NIX[JA؄j52v&zd쎫t͌mPv`pN33^`ze⧃o`0uПL_9b[mV2{T3MBrQ NLHLRB\PA\yڻ\$g#9YM.I3A%6 ;FYn4 &澠vgc[8,Jڅ>xS4Q ~w$"r4&lRK@֒4LXGIv1r#4|!hE{bMK 1 /Xڎ3qam{d`Dl~lcD ">x.8#AȌJE+)RD:dQ.8aJCbFS DF$A 2-*hI3UTx|ױ󝉳AEġvɥ}uv6KE1.dTYs7dT Mڀ5.Z)5ZH# .>. Ojvl'>Cز@XǻbUs|\9Bzcrqޟ ':6~#HB$ScYYlC&QIy':7 C/%H fɗHb _}! R.]튯.GلTX)aR_}Y"AM͒J{*v:0$"4D J Q{Ad5L[!)WyMщypRn"Px"dV5|DslKuSE>Op>5dMv\!]"0+;S$+rSϯ$V'jGKN*!3h`>gęPGUs.ĝ^oYzzZV 9kLU癗ZcoV$TSVoD5&$n) ٮRQ kc̏ b ؿD?<& f۫'BQ)J9;zWd'?]'>^x9I`~?Bcĥ&-SO+,nRqzQK& G\$zVKܚg$0)isp4pʬ+]٤^Mpf^mf:ouf1[kEXj/d)MQ\&A3k#Pd <|]KQU"US9pJ kTԅ f+'6IPF)=,()5!;D)Q- 1ZJqk3$fWsB|Z/un[ _ Q=iT/ 'MC.>Nw[3(䠄!*e(B"IҐ* ՒqLXx p5ߴB|캱Uy}zhP^4ltn{v3\<_BTY$e>04=Zj#Z&g[k9.<%jt9BrY#TeVGjk%Nh*x-*U<^Y@sд‚;X*CI),NbYe)d6( DfC(RYJNKD7Q89 kq0pWYZFj)uoL+}0per*K) \}p(miǝz{; n5.Oa8Ǝj,ZE퍆L }{>M}?9mE*|yVa0Fiuc&8 }B/jѮK1Gc|(']|̛S nMujJu |߾wQNL2BQalQB*@T^kf!&,6ȂA~A^ZNnG;)I✻JX#T[6τP;,>`̂.8#qz˻6-ۭݛi7cp:V1#p ޿oƣihDzN9&(0۷[),,䲸%5JI /7h|Ј\e?CiWYJ^[+)C:KWYg݇ zUҨWCORQgk0\=N\F^'-g/WrҰG-pSA18agu`weǓ<-f-Q8kb|vr*uY^@T,+l-*+y Ф[&i'hRpuvm,'j>g}Xͅ~?~u>˾TxJΎe@lN9^II:yS9qF/g;`S4$5/iipj8OzǘIzq9 ktb򅹎*?䀇;ͦI\ϔLH ^%"i#(UѢ"D!krAXO(u@| Gʪ_ZY74dȔQ$X)\@h9憠b D[LB^fs-M)84R9 O*J[hA9M Y8#++Y2S[EZ6WE%qɱ)FjgpǑ^jcYBS,že)~?xA>Qy;;pkȟ<h:(^ uO(ssmqg7C^OR4__Z4'y>k]sJ`dZ=d^ۺw%kҐ5K{ܽ4ֹfei?cs4_V2?Z&_P.OOg|o#ٯk:sz:\#|_ gjې{d6c:uzobnj璷hN!4P{$MͿZ`ҌD|?f﵁9M_OOOW'*E(ʻa~>/K." Gcٻ޸$URw{M6=lA4pvR }IKF#4iE6zU:5O>/9\edƯ>]C?*ZI$a)4R LԽ}1܈\Oȏ Jo{}++U>r-Kߝӱpp]//(Ɵ1~ߧU}5DȃX-E1J|CKF?]dP)Fm|¶A-JN= -S#lU'i+ɶSIֳTf07ʦ~R'U*ظg|g@Ҟe7+B|r9fX헞 l| #~w3(›5&*i[IzvWʊ' B䊖|Q:R,eL1k7-{5^jG|hP LJWR5eTZSʹ`bсvIvBT\aTg=?t^ro@y Î+-qo&@8 Ϲx!;BE ۣjejɷ1)c&Ě5Yo?jTxrs);UGTo7&VuaH/'g-!duHkd&];!`B ēQ@I3(I5'dW^0ML0C 89i yaR8!P jFrfb]ۑ a8d#bك~Иa*Aل0m*O7w⣜ );!( cB fnBZa ୰z`lHƹ9qM<0XuBLv<@RDEeP.0!Ж* "lX1iDF4Q0*OtP5妃cڴѝ7˜$ `EW)(2fTO adBCKu`Np"8j5B6˚!*EM[5&H(XW(G^:(f%a0F7/E-Yu =9kDv<9UrYMF0$Z'u D]GH롉0u{,XpGk/wC}N$)VgvpZG$/u/!)dP gBYOT[3M)&jcb\1 #IGQ҅4K xQ1H dTvh! &ÉUTVqCQ{0J@G<:(}Ҁ9g0ۧAFs5OP8?FVgdsQj4H% ʷd68%gm8 %}ի?..$]9hA$Y6qht7k5ɋiýP)ji^Cu|*U;?fFmxf:' 7)o~0|VRY su~rWӓ 1-fP^\gfWQG8zYp }N$va}J.HJMGF3&1^pxǁd =!"} chL&@s$#-ֆ088.tޠhx. jwD \1̈́|yi: 0^PFqKIL%3 TY1o+[/Ġ426ydC' YCkV|7/dɹlݾjWs/HLRZrP̙:yr#GᨌDeTJ~Pyݎ@'@COTFt5ȰJol#N=LB)u3%LV[c\c0#F,Rp#+"$A$B㫇 G0c#YgFPi "w2m@T  h$uHkő6G$o\ .HKz=$.wQ\6Vh0"/xg+xe E!H4&@`t*ZIiDv %2؈bʋN#xwQtK<aRPG\dq/S>X= HN⬐tɓqDwrjt=q’z 6 `#堒*at^)7VcL*Ai3YqQ&I</uZ4 Cz(j۲ۇ ?b-6xbo} /\#NjcRqkaNQLbwQ'RމNjcQyiVEX1MFrcNi. AE'1h)cc;Mm95>WcU0>ZLl g#RV< |qq}R%WA:ۛ_a,Wa]e* $Ŧ}*?UX|A5 4W'.#᪸6wߑi{Co u ~'4+ТߕɺB7etU}#)v%ZE )SS(L2{>Нud=z9Z[NΗOU2NSl6Z*md5_Y`!/ez׺W]o;de,_}!LfH!Xy \RX8|wm6l% m {)Cx5ZV՘!=5&CͨweF*I<)C{_rzO- ŊkB;#K3|g~y]$ +FeaKخ?>BGt*˳80`sXQ10a1L ڬDSTvR iAz3p{ԍ$$C:0mw  ;򯸙@M2@MMUpZ3?]q}Id";>{:]YEbZv%ϙ6)=`!2lyÛWI7" ~/o{ϤzW 0F,?P)It0&Jl$C/G:`r+?o30IՊ4jLpk+QuGwU?˻@YfYsULll;(O(k51΁%=J-poL.i4/FjBsѨ݆h`! dWo#Bv1i^c`;/I g}ouHQE@bUuꜪP!CeXЄrf f}j2Zc+v[TJRO5c>tJsiȕ,x\$'|ɠ3V &; ,#f{ՁOϾH@AP]Jo,JH:RBq` zxѲ*ӽjlSZŒS Ho~:Ukqx(3˗CW(iQGPdke"wBs:l r,K6 :_;f1cS7)xDqR,h/!O9JdLkODg cĈ i<@Ҷpfwث|du >,pIG$G# (  |TYsG.$Mu?}fWch{y-sk4I}L:.hn+:id%wC)zW}q$xӆo2_oK:wK3~, s/0i^ hx>U=jrZiTPlr>l45Y$(bc D;Hޛȍ(0[ztĮ/Q<'ۦDm^K<Ȯk= Q{]$5x`G&I )xkfe *] c > qS ft^Բ7_Knq~{{~_]4<Rg;;X8 aޕO3^4<~V¼iWju 4zSiO^Rc;^ qWy2Ӵ=PTO>41J#,Ҫjvmܟ]ۭlRĩX.cPT̆ࢍ9(ɮ3Ԛ;oU6(&T[eۤRҐ2mJ+.} {ˋcȌd.3B[iEEjT(ES>"C'a P%A%ZgZUU&V" ͟;atԟ?.Zi<>?-iNGx: $Λ|JC{L?.SZp&+^>RkF9'j O$Ahʢ)o( R3^XdR4u[`A5)f`V{idd,O1ѭyYAbن1ۓ8P~c3!Z&ם"V?9d՗ -=~2/D1yI5h; :zfTVd2Ɂz}]dhr{hHdJl.EY9Z:ALð62V3g=2*հZ ue,=z0I6.@ONl8]7 OoZF@ ySh2psc|4xЙ-QfE= ) lJmV^h&chP6)s(1:+#v5s#Ԯ;EaQ3'ޅ$eIӘCt7t5# 0Q˨ %~uČg-1CHE aML,h0p>f%9&\euV%⢭d89"X-s@N6O-8DmrP[vt9Wo;: ;mА%C[Hc7WMt?|Z[DVұ 4SذzW߶$f]d~h7X4ab2."FtW ,^uPoOƗ˓E⟏ Mo՝[K$q;&&ܳAsbx&ЦGЍc5%ڛ`3׼wޑ{6y~"½pvdGS 0e}$n]/d2hLMNȱx9 R;\+UW S"`ઘyAZW WVaiwfL(W>m|^T[; 2Ҫ3 ~fpIk8w4zryqz6T+n0|z~N,\ h٦o~/o@&y%iGd ",4^7Ғ~ps1pM(8x2xӮtC5`pw &YwjR_7\ ?q H9h뭟7 CRf`Ve;o(V.9˙~@pU `TPX+^+ꛁ+aӋ7K/*8K>\/ W־K飬,#Jpms`W`ZsUU^#\)\?f<6pp&-`kQYv^tb8^Mx(\TQ3c<!*2eKΨ!(1*2< fIz m4F 0tG 09)L&NJjo.Wԡ]6L[;osٳKgb e2$'>$ODò۠YQ$z*w޾A3V4|ʚ( K`7T=7/oO3,zuւʾ&Zh-{MDu1jgh8rwħk-vWvy;/mpHL.suu!%֐մ({ X9͞_UCȈAD,Ѷsq ]p iN7.Wz;E$wj8b4 \ ExZ˸_QNb1QPPNl@kS@~*QBe(f4TTx(dv:Ӫz:iV\!1'2h@ # TSId)#^ܿ d,(qrr8˲r\O~0搛Ѹ:˽ڻd^NE\T3&d/e/.doWn4s<$}8D?eN'iEǓlG|̏)r$]~evJLqs!x]ﯣS.oMy?4!O}Nۺ'mS.~~8R~jY r$E=p\OiwrϨX[._?)G }DΔL[z,1'49//OTy`C\ڰoMe?;>W\TMzs||n%r*MeE{`5ba^:,||2ͯ?d?@flnfW-Vj~ XvXْɺ98,jznſrj>ϦGieaL„] B¤3ώÆs9ޓ.`<߮#[pevwᘰ.f{ôa3&>n%a=ţyfa(!<@zVo}jNw Ca$DMZ' Qkc(ޱhEIrHT[yr8'ydȣBoBsa>葙묦DLrX 9$!YJzs֩'[dxŎW?yyF2ZN b5&(T82rxQ2,QNIāU0QŸu@y~SN]{Pw#z&F⤭x *O*d2 BP&}ɨ[/ޢTu -߄eSt4!D`hZiOb­I*$qBVj QsiԚBH"q.9Vw, 0+FΞ֝6,LI(OGo31:Eޱ[ϗޫIAІHo6jHIZJ%FM3zLxx>_1,h BriL&YX@Xpo\Xq*xYģl ΣI1(K3˄tA$E]Bвc$4r9mob,S@\=BLȤ*WT#+RjN8#*0u˹aU0mT|F_195m uL{&ǜˣ } vZrR*^TxPg ZRpUbUPY-EA@FS~`Uzt0G}d4P(SP<D9v'nؚ2^~h˿sn_};6Zq؛Ḽ/~@bzo"z:|?}LC%6h{r5C< `CE<ۺv o|SLf%4Z('˥6bUh%2ƣXF{S;tNm|.}BLEbdsAl{&҉8+Ѳ@Yo01BJ&fq9Ks`,KuN'i)Xiv1r *TjbP΋tY`%MeD#$&NR+V驅 9mN(R\vɰC?|y^c҂hM'sL|tr>9&q`ysչj!^)RI*WMqK. v75:DFzo̸4-%@ {z@N̅}I45{he֐ۘi'\ìI٬w֟6 AD5 $%%5X>ͳwQMDXg-qBlXja ^*hQnzij(|~k9j**A}7HW:=]BkpGU]@iX+K>R3%s=doQ'U`86􃗃;δJhp`p .5I\ʼB 2b % Ye24;$QH4CڌT9֤9+(-*WĪ&KdtG,6,6{ʻc"CQCD2y4$"Q>Yɩ̝ljFet !9GL@0PDo:$XLJt *J Űf,FnXӅ8P] B£pm.逞~`ھ̀ чA:56 FQ M[BhvfBfĵ$=v'ː=pgYmrN :mGML(J>hSXc#gvaX11OEkCv(;!حUJy.Ә|$oQ(3L0*[1IFIsd-"C*Њ0}xV0EUaۂg~4V/U:e@MWH JR |J-InK̶<FV.oO 9Sna͌I4 CvI:%Ыގ^LS/gtxy*'=@3ϗ}ՄTaA+%~A`&'H͒x+qdVYyNك(:V> k$ : )0o&1/_)QRx"d6-3ɤ}Y 5s\:,'t}{ȤD󖣛fw^q|.Gǝi9I SWhhBSL ɣ]F,h=礨S-W># Vf)͙DALI`(#jGs@P(Ob;5OsZ}œcS~[<𶥐ZOM_~&a,Mqv}&ەR/{dz,D .H`(KB+UHŸqVR;ŏ$dM2U8F$zOD:N8Ń$%-ΎK<A`t&AU x5ĜCJJbFsZ+8dh}^M3{mRml}̸WrŰkͻ;LЇLϢe^fG3я&!x]_}«WM?<7ɟ`\OVyqmF^jg(5ep,ߥG XYs_^W z;APqrj=$븫}lYeò7^o$ bwu{|Tu+k&X\[`W{M~g6 ˂Q7?7>QS:Ny_=ԟ^:nQ/Oe^=Wx>t4ff "(2 Âmuޫ1s)\}TJxb+'99Mg}Z6+H/D%&z.}2DK#w6QSR$*o#IBvq߻{p$ Hno?d _miQKJ!)mR3 9OUU]9فUζθ'PlN5viw}l~וY[tzV2=e*tzbd=P) vRu9 ܇t3OAU*cO8Xe `~bk}}}!!$$\ 'P~rah@cf%\`2O? ]m0t6(h KyW{' Q!v&[aST#8گ0[\bb7]_PRwmq!2ZWŤ֯0m@iL1 9U?#IPHLjH,2n Y2+Ʊ* # nXx (p9;quc)_wWm Wogvn_}~x q.W$Th-::7轗 ˌ֓4(d7s%@dǹA 0(2BhG*}Nikp.wYO@@hbA| mۦhNqnW ,KÕ1[K#YEmϪ:<BNWE]!?風*pўȇDŐy펙ߜa}69Bb ,&ˬpNAhɐ@̺Q̡UE,#$ŎS2 AJMB;]@ iۇgGoŦkAmi׳n~ʴӭݎL<2,qY}.Khi9'vՒ;) 1ፊãZ*J2\O_9\ks/ҧ?[1cҤ("aUW^pJG*]-JD(HM`_'+sEnɴn +[Xw<e Xc=sQ) ms4!oiAkt?li'Ac3<^h{/O(7\@|2jQ_}&r3\jdQiyE/ o4kOj|wM6d`?ȫ˺wB Q1{/_GO ~!aL`}$9pǵ4ߏ#~/ޚLbΌdzkT /rl6Frz}=jM{\99hIi.= ;KV&f$~7)~cվi&ҽw_t?_ LMgnKgvسi/}l>0'a%c縬)OmpD;㞋P+M 40|<9h.:P! R"3S`NV[HZGP2s @a<̇=V;[,'Z.D}Cv: Y)jC:N/ʁjV"]+rĘJ*d*y4/B:a:a-V^(HbEN \!dfR9KڢNG~A]uQ_J(SPORf2P9'mVmUPXmGqhZq}9/ykU~AߛP/[ ТhcO GadY%.`|!sY(|N/ 6Z$ IӓAΞ{.?[8pH*).0*̀2P}ԬZ3ϊVHa(+A e( <ǤQ A0&dɼ-Ykpvy+fKO&*Дa+ VRiY&_icL<@"QG"D1d)yeek4H 1ZR$CV VkVti vqno( ux:f S 0)m% EfoĨ* )ݑS. 8cr O FYWX% PYR WIQ(eJO:uӤnYPͻYm@88kvQg?]Ag~?}: 0eH%v\#?~{ZaR8 %\ Y|YYFlY/V8ƍ~uB9 Zh!I{wIMlͺU|\ꋫE ixSstszϓ7,jywtY\]. td}usjd᳟;,9Rg#O lyy6Z,9tUw'x{e .rsհ&3!` jy >T9T)m88K| e(>+C_gU?XeVk~d]$i!(͂pEB߾#OExۘT!HEnJ 61V BG.T˔%䤕&=6L{l;uiJZܥx\E<.p}Z]rGtE5 ]\BWʮ1 ҕZ .PaE ӏ+9kBn<ÎjT#GӋ5>No\j.UMUR%*7ysvH$GjEDwTTrRkfGoFVcZckzǂ|d[U"73ߌ//&R0VB(\Ugd*sU**B6AFQޟiVjAth]+B`+,h#BN+ɀ1u" tut8ѽ+ 7tEh%te|W]= ]ixNr `7tEp ]ڇ>p"z0O %C sZBO8\HW(nm;^dv]nuoc훆@KKw5aRcRezZ_iՇzd`ןxS}h-w]hNҢUWD숳rcX܂q[vo7 d46=r<\+BӄXҮ4}:4턶-VtJ8AS{vQGК#m:vރ@Wz.GtR+u_Zu" :AR9 =+ 7t%~Z+B+:]J5hWHWR)мGtIaqi/wCM_#eF-WdɖdiY)v,>|X>v }7]%))ocRS$'WI\NKJd W/8G$'W ''-M$%U-\@O@`U*I+t pJb`+ |A Iʼ=u졐\R!Nh'~&q>3I+E$j/rTfOd;Ht12u5Ƅ+`ϛn%?HXq"'q>NҊ{e0aZ 5?!҂J{IDNRU8p%vzq{_>W 絗 Oc@'%m\=JpԡO$`'WI\O6@Jx W/$.>JKu2pT %H7KW|MO*vܹ/|/rSI)_itJ} M{d*'yy,-mT|n9AIc7RtSAUpQ WW W9v՟'q"q񠑙9Ǐ&onyeRA(1xf2t e*HI*f_pF|A{BR.6ߥQ5Wz/Bf3ME 7ݼ{b ;P)?̞l=djqq6fO}8Ӱ^]RK7;Jr.eQ{~NHL;1Oy0=U8dDx尌J"GC8&LF-_]Tܦ蠻98C!^ҳU*]xk&E=֫EV݃'Y*7f_,N?~V@leA?nY-uONfPqN@qltf0H&^_Sg;'kr{z+}ZUt;oA[d2*0g㮖CZ>Q*[:v9‘._Lq%hٌކIJhaSuHeDzKR~;}@v2(kf?59BQ04L`D$| (9Ix-PI)b|z/s+wZ,Rg2DO&SDc%%<%b"&-,xGI4 Gj&G89L{6hx,q P#ȯ%mXs8X_Wi r';΍ uh [9@(vEŞYP@ebZS ~)ւ8+,vYo[Hr)`^q;9CJZWnҦU]vli`>K;N03(#ŌY3XGXf ѭƎ{;mX:`ӳX^ǒop@ٖm\D?uÒfCa>Rq>ȇktf#@m 6bE 8D HI=`  1:bDLjF,PY+ O^jʈhA #!2Xy$R*MǭRָ L/$AߎؼaVry_[yU45{>?Q q8mW[c\c0~#F,RR8pDH8/}X../_yB;fF+3p3geY@wCʟ&Jy?}뤹2b\ E 5GeE띈ebnugQxQ@\G], g. Cs X?GЃ|_prNgJzA8.' %.b? | A1:#X3o:o (yai|T Jq`so/CMqe;?"1 ;גg`YUti D}8_4*J9 ￿[{o)Rp% H3IA MR#s)ͨ; z_Jkn_ h j_𽋾_V[>VVWo&-zFsR&]U:i;`9ڳ*C4a2$l )S%,Gl7+ȷo)VCT }`gܘdŔY.<CEg$CXʇ+u@+56E<$~c\,`;̢k_0,*n6zfe?),Sj).,̄]"̤VOfo G\-s`4_/C}V"0z\I] &ƴ6"sAse{N%AIzhQu|sw$݆!HiPps֖YLTJ8*D, :rm t&{scFO)XȔa>#H_T")[yS'"-%)繊Ei(0t(DXڕXh)QR&@7C?Qߐ*5(K)OgV= Gg5ӽZgi'I@r=_ӱd~oAl]2v}֢2G텧N1w #QЁ@ya$eU ;-Q1䄥,>eFS:J9xe˳8A94mPK $j2Ze)m)Ց( znWxRǞR2a9?)PȬ.DBxT%KJ9r5mז>IOj"p**% s!csc!#:xlviaN#!D!e޵ٿ"S`~wc8q%3H-)Zo )_ZHwpzǀ;^cEf[l2,$HX/j/jWBS(X䐢A+F@-S\DH4dVeɌ1YNbM$N64#cg/hW'?ߴ)<ѤS9m>|v:U9f؝OR7xzOo.f'C5x@Є%W9{xնa9F59$&?O>y:?Y Kt9c_,tj>'0 !|&Ѵoڗ}|Mӝ_pgkͻ Fu $Z GA rكlw%V!B`-EPinn{ݑnnR5Zqk?^Uq^zv-{B] ln7DΗ>xK.7FmVwklJn:+ƿuHy NѺE _u4<#17w+ۻ,=ckcu4u+;5|rI"wl:q=fzΓC>lޭw_sn[̍Vzj.y컳ri kVjS(tޡ, +m/])<+m;vRUމ6t"muCS(%J p8zW*T"gIs21.(JQ2XD)}RSZ:B,n7eTp'&azrDžs-6ʁ)O4sFB=(>35#-j$V5Ԥu9! .$9Ҽ+&5/ d$Vlu{r 8 5]X͌`i=\wE>y-=y/[6U(WޟX*4uX&zfUΘX}v$ZrMZ_39=4 @$0%PZؘS 79 3amXMV iƱTf vso 2R;xvߝxV{58 kIky׷϶1fx_j.m3L_Jimq2?m_Hi2Ajf>|hKjTpYc `x8มh?(ݟJ&\k|.i9ThK/IJ8τy؟?AY}ڠ7*;i͛wPZw1 AL%-^mn~E]?u5Rc&"evMԟӆB2~~f9PaP&Z," [R8Xb8ui|B /1JBP*Uʒ I%LQ !2 f*jXj1CZ+F"Uvl5q[Fۚ{5]9EYW 򅘥QD.3˥[%ߦzz.=6ݐ&hsO<9Yj>{"öGEm%Utj q0 4 2+x50MO:c&1b"(|UȌR>KX)(O:v;O Hv"m5q$~w޴GxQyJfdDmt\>[hR:^ny-kWt.yUs(&sjv_Zr'efhT,R9oge L9h*:}Єͼ$<6wE6KhʳH.O 6 yY *SJ3JAb)ݝy&q)]1keHn{ugt2.ʖݰwᐓ۵%_}+?+EځwK?wi:{K'&?voIU7y?qkr&vw1'KyGF?'w'F r'_?1a唖߽xK2 ñjp\cpҷ<4ʲt+%rUGgyOǮ=jxoo/lNeP5tjoW%Ě޲2~.U`5SߡZϪ󲋕jS8<=NN>gۮkĖΒ,ǫbNjoΐ=_=|sLWv%y }mW9o&֌tӽlJn0t^b1m#o7̮/@~O}fvJ_K;޺M MiZSBq X!X rqgr Aы'I砛z~J[!g༬N(zDZ= dƣO|أG5O7WiRI',**7G9v);ULfw`҃.H`2aHJ֝`V,X+V 9bLLs~#0l@%8TLkN>!h|wGf9ڲm"( FIK +HBC4Ĥr !SYoO!FL8!IO#dA&xBc%WjGWC{DF$i~5y?пo-;ˤ]Ř sHTBp@Y{ LH\J1L2a#˭f (ߞvv*\_KT(R 2<+^sbrI:n,B S"|JX&N4!A1%fI2tU@;ݪݱf\)ءt/9G.%3 zVs恥V m1'np wmL |}+p჌R :(C +Zp!$A+XڊjUZzܧM7o, A(ŠE%Mm̨ '`!14&5e~[GѫͲj:F6˶CU|4s&cQ'QI% HbEhbKH:p$MkWDHf&_z_5∞qzB"ꐱ0$AL*%!x mS :6G=cMZWӐc*[Aoeb 8eNiSe7v HMjMa9WF=O͛,fq-W I t7sl! k/=I1#zTG !)[PʦK\; C(3J, cF"d#d08+@ǖ ZW Y(;A q,H Xώrɳ_ˌ([|FO[l}{:y# n/X?~vh-iۡW.0IS*RO'vVkp=~{ˎNģ'5|PjJk/jnwZy6ChoM%/|o9|wux7p 3*ICQ}wZ#jm՜{IC'x@lTsZ*\{3ӧxuWxKuf[-%-e{/Wn4ßoßυ-eZ9n5J W|^B\O؁B$@2VJ v3E])|/v)U7ܭ4ԝ H3oҬtyJ3G>=WgvE9&'J񲵊g2PF,/8!Y\)Jgb~Iz{Ac[cK;{f=0a}*VkH_ƶ Z}l o,чG8ŘJP*pYo誠]S*(=#+^&;x;ПC [UAkUAtxJpzEW̕ ]􅮀VPtUPryGHW h 5]7 Z! JyHW#SW$MUAH J}HWJtXLjWwb71>5' GC5:OMu\/>+W ڟ&,h%w*(O=գ+GtG]\ @W4//%(a6+M{|:mᦊ,e妳d9U`:-˥P?04hn`"!M|l:,qSlEQKd@pVK10'eɭ ? #o8$ttx-.yʼn\ZUs-ikk[K S7Ϣ^|Lr:, ];]mk YMWCoy??,]xM9RٶQtetҊB{CW.'}+D;]L1Ջ7F;waF.^^D`k'ՇA4[z5KMمի_z5s~tE{p_꫊go˟R>W-N5hw]RGMB Bʫt`2, -.L!%i~ptl',74?FI~ Y^Īƺx{f nqkh)qWMi*8T}ӚcET]ܱOMe/s=#PQD(cd WAtf^7UuF;9oF3t#˪K] OM OGb^`{1NM Mp3ir$SPD=W4(qTW_/3H ^:Zfwm3L6toBUIY(]p)k))x$hI֐H.[#4-pWKF2vUR q%.IVg8:?CwkAr U%sVv ( AԌD.4Pv0eR11i5eG -,UwAvԡ`$vqhvh%hq0^PH @:YJhf! ё]H %z-& ʈӈBieyX#L%X=b #92(hhک* mxjWF2a םN%T|U*bJ<̀dP*, kme­ ,b-Y;eP!I ZQ>`DR ̓ @s8e`+X"5R> cޓ.m(BE&cF.E((t(w!\v,6]5Ke2rܕSc$ "`ր* MHp 2RP%X Jhf4^΅qĀ᙮KnA. s-Bl[3[#ĝ03D6ӈ/}N,tHԈ3#>`#3 y 3` ? uƟwd?Rzs>^bE\"AaCtiS:‹9Hb2/yr +.OЂnD12D=,A۾EFhwoq<EHz c|@({!=$6pЖ 4fǺi좟0hm@yН)CeG]7amI!`fh5.EȀ ݑDdD!Í哆a9wv2Aa8buLS]:їnz/`zX+'}x#4Iwh@fַxۉad3{]AtP,B04 io[%a7h a|^y;Bt{Ռ783iS#Dηтf1zk|?aC`-ΔnMY:gF<ֶwbԷH[)kjb4[ a9P8iQ@y'#_2Y/E)łx@6sЏŌw1sDClSǛۙp8RHBuYQA|ydL dGd P>.FOd?-ӮAi\.R[xCBwT}5!5(R9 6 &tX~@jA.f|1Hv)nL`Ȓ1p=ig0 8K> # ux@\d(g'v)Ksax0s;Yho$MH:0ķх˟zRo $a|8y9o1tc. 9&+1Gcg^0j\z8Ogq(߄7 M6=>Etw$?ymϩ 5]׏ ^5Ӧܟ޾ؼw7!Ͷ?G rxD^__}.S?4Wxswu>@?yooI?ݽ|ںta,_g++~ƚ[e|572rKChCqǻ/ڸkfw;?NxBݙܸ.($YpJr@[H S;u l$k N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@ZpbEN 34/ j}5XeZ(E N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@z@v |Kr`\:z'T:N5:KFkH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R':VɝfN vI('ԆT(:6'N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R'z@G5k(n^ᣦqpÁOq?wn0dLQqF+Ƹ~3j\Zq3|읜k,P;Xe*c,8\"roR\WT1AXp]T jqUڢZ!b \`sˍbAV}b{%WUq?pwuͮ7mۍw[ޞ@? xP#;lo0P߼z!5MǻWm3@Ms2|o-pnWϿa[onisջwÞ}صy _xyIϼ߿yEԵ]3v7ikJ,JS׾8yj?o>ͷ{[smG'k\$i_ ټ˷ڼF? 55Y!Gys~[&7ݡb)q UW 8CgxO+KA Xm~U*U\ u= cj<, PRs̚L]S4? ?^Ӱ6\?loCcSsdžBs>4w]OG+_]ƻ淡~FcY?z:?+4 SzSL9 Kc>]PGIzy>i>vrپЫk7CSٵ4.9wT'L;tacU©LwMhKl 4.tČら`Yd4,e) ׋Yeҳh"@Q6ll rdf;dm=,<\R0jcӬ2 1ͷ0%IQ%F1z6W+t{uŕt\gv0ޛ%[3.}hjڸy*s])+zi[Y/a<W7O+%z[;XeȊ9䦳9 WU+Vjp\y()ڴeUTɊ*l+\An Z_]WQ [+%'fzWFWkU$IRtW뗾[oZr *cҝ51P]ϵϟ5U]e΀pza?!P3&̟P덩}Οk?cܯQŒ~QŇKLwi^,LCmc*SE5bD$\AF XntRpjsWP|R\}5rzwa"Oߢ#7p \Hbpr\\=Xe) W咓+V[?2R\W+, jR:5 TEb\T j^<W9 q,uiL4g!u|qy VJq,LkЮ]n|U^fVφgU {,k/˩6gIq,mie^=u+3>80N[jCmۮoi{}HtL9:Zn| ?|~qn[_tv6b_}F9n?%e42)75 ^.6N-'C?i=W1b,v)/c\`A֌vg-(ܒ,"@mEVtGy9GZ 8/^l( -v\AeٻNq\\L̂p˵bZW}t*uG+•f 'wCҸ'7/]RBSY~]o YG/W,7-v\A3AqB\\pʉXRpjC$ֈ`ALrpr+Vv\ʬqEN+!ˍbAVSj&")&l rҕ կ]ʤq):Ib r[(\P}2*#)V+W?Tk~K|lZ?c7]:w̅5!{GKlOFP!ȉX6͡Uva gMfAb9_WfyWeXe-W].-q5OYrY<_8!\S*[D3pW/zy OՎ+yYjC0UFZ! \ApQ Kf=Yj[HqF\yo(xAbENtъIYwUZ!>AbINtrs+ Kۘڊ|4A\ rvj]kW[ q,Wl+\ڐj$ֈdW׮=kgI5=.ǬINaY7b @ٽ g`YѰ 'a$&a)֞Ѱʬ[hbkinpEv&W2=9l%Igg';!'ӬU*)WR płm={kfkktzpEz0dmpj`g6͓>9Om,jb)^|"Il\\/Wlb1*V+G+W|=yrs+ujLZ!<J+'׉]Ny*I5*D ʉXY+V_b]: (R+΂+tyj YqB\Ř-płC+KY XmUVSqu\%Keo庚-L }{Y:@ˌ#fMwԱ~c%t2g;IP7$fVYtUfqCN8/W X1b>Hv\ʬ{^kUI-zN1-bph2艢W]/ 2˲qǰt28On\x~ڼy*KeUzilbpXf+.׎+V7WcqnzR䒓+VXeV\W>g \Ap>yTbHͶv\W+U(7W,8\Ֆ++V+>HԊMV Xm)㊽V׈XJVXp+[ B-q*S\WK$͜%WBYVve3A=j S(kV'S*RIٴW@{VP=3wa}<ˉfLIkmIE.nh!]$lM0 LuCIvE&)$[)[I8 i^oiFekZ_ρכ CҖ-Π9~Zn;Yܳt6Z.Vb< ?ӽM].F䷓)Fb~SPnq}:u{&eG3Sw&>JA4\{ !HN4",!F0QI)V c΁+727M৯mhti!#2DH^"9̰4N+!@$V <vP܏aJv4걕ucLJ )`nAh`I0;eF00`rM[) 08`(O2A )ZyBP`%*:_GEjW adxuC#* "z2͓%Qa`@GTڐӗHص,K4ϗ΋?OnV#QHn҂ZL)phkHW0)N]u*I+=y &CcJ3l0-@Ѣ&/GqC[s0 n>!mnYƯY1:;IDʼL!QΤ`& A-qɀԒtzL@^k脙MϠq Nzy<}V=Oozr@?+V:0}K@:)5(ii!t0)gw_УT_g~WsV}yy(S%]ۦѧpt9_Ocgg&;foʟ^ӷE;dW0>lo8m+yab: rWz0_o?)ZA,'w5.0\(J'kP B7pKEiZ&2%9b;WhN%ZRuzp:o{rSbp`3DAo?ƅ8ds:}a'g_ҰشΒ2%Y{ۆGVvϖtݽˍX{F(JTي5(n % /?}Q@:O,lxZ7#IJnV5ǻK7{65c'`pk0̕SiNQc4BcVG!% QqRǘ'QiB{bZCJ<= uQ Ƶeo;9gw\2*:Yi)o(xΝ P zv 6U+kT0EpMZRm5i_sARUj4v<;Ѯȿ}Ki@ƒI|dePg ub]{6t2/IuYߧꤾ=(k{Qzw; M瓔%%]HjD.4>jH(0 5D+ j1dA!"55͝Rࣞu ෌\2#@I7) L(偵gEC%DJ,VLZ"$"MΑC*% M[ֳR.kwW5g n*>*OVkTMhq^,srWt{0O5P#"X- G+ZX46nj1& =jcR]ybWQps-fD@TfFLP5mAxqsH69eIƹN@:Ca4:G{Q̍ϯE;ZC]CgL&ɱ:3h}G$+dU d^p,6ybecC=0ΐ`C @u$B6~~^,@{,~ɪTT[1Mƀ!e&jab\1#/9ȱg|s@F~'RILa)L4 1ܫ5sa$VZ,uz;(ERcib yk0}wu{ȟDF6Io`!j͍0^N"!q %BhoU rԔ}a0.HYLgHkXb#p@i AiLo𳆡^Lq$\@7VѸ%V6W!|>1xX[xY!^9Lk ss50= $:

tL>Rwޙmߜ $L!L܇65snٍ٬-K[{:~;mL>/cFr5t.ӳlTx ۵@zyunc2J\O:tݞ)l#15*Cz Z{1[,} ^Evtqi=SnE^B{n<IkpYu.K>'0Ⱦ%4h|9ڰd_Ͱ[C+K+n_ `SP$q襦Ā}4<)c";MPFСQN΁zpm/PMGF3&1^pxa"{&CDT(X{)i%b"&-,xGI4 GZ&Gs;9/AF @&=cvK16e0Ç4+owy&;ӪA,x!sBŁJ/Q$1-bE,7PeQy[Qފ M4Eаa'f[l8 pXRGo$xP$ܸOeQZxe@b_%WK;wQ)"|L#2.7< W^OL* {H ~!<7xl1):)Kn bO\,Q- ۚY'pgzk9~Q?o{͸p9 F4=4Z4:BO:)|u(tKy4)Mg̫-n:H RR^fJt@IX33I0tI 5}إ$+ S3č6b&sؘ B3Jz tS3a;*Hk^"IK"IIű',J}8g|3r&ٹd%!E=_|DX"0-rM3$=#R}Bȁey؟Lnl;}&fXT&+|xm~0L+)$ z tZ<1 (kXy*, ]// …iVU6;8x"L`PHb təBb% D<K'^MJ;mF4Ip#h ~;T(T޵u#r/o#û:EvS4M`li%9MP$dKvJb1 93$ ň>< yƠ9vKn~h3n~( +vH֪<DZ˂/:}x|g@F+.刂w;ܤ#cLFF2%РU*%Pd%. ',"24\X,teЂ J]@"]GCW@ЕA+ 4J i% !]4e82h J!]@b% ޕy4F д;]iDW/80c+ CvʾOG;aBZeh6i|S L6zN^. KSΗ}뎳^Q-$Q0ȁf3rzfɢhGs1α=pr%[- bGD0 \͙FQG%]Hӊ 1y"2pe42$jJ:^u4!ljtOѕ^Ǒ JzЕLt^N(/,+WXʠU"t(DW/p]ECW.A"2hANW%E^ ]&ѕ,T4tR`ЕApteP]D⁘A P ] b+]2yW/CTDtexA Wh Z~ݠ -,Փױ ѕ HDCW.DCW-ӕAIK+y0%zT\\OQuWP]6Շ\?Yj`'az%\+jr8Ҋtzu!WKi `ţ+ 0 Z"C+*vRWCRҽ M.gC+ɥJX qAW*ն\Q"+  \c+Z2 +TLt ̣+ ٓRB++%ҕJ(p$TOWZ:pꁂIya2f߽~=ȺA1z*f-'GKwe9geV<_i.aldҤYӎJiQ]6stzc~,Y;\e zɫbX;gqoJwu7E^O̵U{_|ZTt' Fgb#n bevDn*vDqA;?ou|1+m&!2kv;ǗVWoLgMa:~O7bpTDܾgW>TuJcn\)Yߎfێ]7kGףw&nٛgr'Cb!dDNJNza`)vz~gvt/Xذq}~hùȦ1 fla?q|ajQ(7 u&6e#?nR\XTLwbEi@10)FaA|Mnx66PrL:htPw]i/v^M&םkʹ!u˦٧1+٨|zk-fͮ5Peucһn}7?uv߻EWHVKY9B*sE˲lZ~eXm@n<)O yqLkM7xR?SdȧMKR`O|lCB23 m}I0\{+/=/&&mj7n֔{g^̴y\2sc$΍IMbj6\HZU(jTpJTP-k*_ŭbX񖴀D!*MS;Fg]vtm3+vMl(T?=6-of˦٨ǷzVu3sێYFې@x+K_15 eQMY9Fe)Q)\LiX( #&V&}!Ƃs?ڑKnmg p3;R~d`h/eL, ;Glrb2Ϗs4OjfjEy_0#B &͂1'5gxu\e|׃c'h.tEFtbQw87vbQU]^AW Zq+,))ABZ*5~?_wh;̻ ^v1B,g,ǥ8S鐂&u^Yq'FjaiIr0G˂ qjI#o_͒;4K+dObK3 O dukÎƋH`B_lCV8$ D*m%E|/ h'<>bbIJ SlMnƵ^lcZlCo bd֦E$;#qhSk֜,oe|zZΎTB\j.Occ{9xbc62[sj0Fojśq; @&hH# UF 9Wtjg|?f۾&r ]LsWSZլߞ1Đ}ґM+Ů4.5]WmnNobPW& X*E(XA+IHU K*IQVEӶ#Z #Rqƙ* 6U{yn`n%}73`z1~&Ίɫ[Q}xӶ漠7mF\۝ ZYlsIߙ7B m%W'5*ne^lmĺ1DzW.1G340A|w4Hn=qLb0t'LDܑLxՓ/% ’c 4H #( G;:G g);Av^`,Ģ9֖d),xitvc4dS`>r\|H*G;(FYc]gFX|VXeh!.Y,NuQ3ZIF&رa|};V?wY S-L*&7ر2V(q.^ۿF/ĐKFp,T\4f1I1 gu,$P)-p1Ӯ5OOa(IJ;2tr12Y, MezL&7dU5XV0I!Ykda=wxTؾcFct|,VRM2 6l!qT4qc CwKɅm=ʹ>Wdb67*o؊c1;/{=lΕEK˺` `'9; kŻֵ۷Fjף[;[KljW3вB D1Բia2ZY[AcEQ.W\zvG}wc6:Og=[.߸̉:<8npA?n\H&o;JV,5WPU,Η\;^K$ =t h* 1; 'X,I1nz)*q(,>?gT/K[*uRc ?V]{x7K`~+܂TX" ?x4%X5+EQ(~֫He^^&+1N'~d$L/(~]}Vp,$੽>c0,+Y4>cui'HGlqv^+x]QxM{Ti*ziX^mJ (C:onNKތc p'th7YjތY*}~%"M:zJhAJı 77r6ȺdPLGvh'T Lz-쥁:J]ʃ6+]m%9DBTمm7WU_¡)e%E˙ਢeIj,QE*4mY *qz#:!T8@Xme%(90a/^I=KQ7\ejOFZU 6YOiy5p,.wmŢ>/fE %2AߔfP[OgR[k7ǜ`-8fce38{_g9z<ͻhЮ'@sH/Ӂ6|$nc=Bw|Z\3Y7J8R ݷbCFhK~+g.Gg* 3Zz¨h^=R{ Y:pOzOJdONc>C :|3,ĖByb\;@ޱR % #*@k* SQ'2Ѹ> I/9odG>Cc B\ ܥ'u sFƚTEYzE HTf DD kQABm3: ]sJ['~% *(ر&(W/3e4*^v*Jv^!Eq^Zf$E\;3yxR/^rKk50f&5FKL(W/-sN1qM%4#i/G%bnHp$aw'7bbk6B\ÛޘxxjbU6)YQgY[uSY7E]bNUߺ%QRb98m)LJ+s @ģ*@سe_jC"V_X[ XlQe*X}\-kcC1)޶ÙA&'?8u/w#WZP|M|av/FЖbaŶ&( 5-Ϥ￟|~+Edn_ݜ~u[Pk;!/hAbJog烻9+6Mf?o'zL˵cd:ʌ"u:o:+O<]4cR"M7ɄFK <{6?ZH{"낲1[}ZX^ւBYCh'񠇴~ܓF!^ѽiI(un7} $匹 svmOHi*4*CT3D!xXJ<B#WG VaxAl^|hlo3Л!O;9FhTiZ"ǕyPv9aUC0FJt[@d@="6c[w8,dՉHLE:y&kD.Vb,j#Zz6xnV?;hX^U4VOs/5M?pA41R~KFxk >EaX)%@@m6K o{oJha;'A4}>܋Yc߿(1#N}&)CH㮧^^ u5 xZ7+铇?k1ׁB'[f촾"# ~MxOIV"PBgnYgR[⧙1-6) IBC@UCPb-pNUʋxAٽzkݣO KG$kLտb6[_z`X1*mrX|5)~ouErqsd?>4/1#a-fҙ%Q*Gs83xt$dR԰oe_\.*. +b(x }d=] 8"P4)PCX%vMz̞3gjvCX6x_`6zۆNn;!I SV(IAuEQ$ݳ`g.y6ăd-q+L_/s\VVEůTD5Kw47J*2#qppL̙T_TɧqH#"&YF ^5 pTa4"C) 4~^m͓H#bdeU`G ԉS S]yI"/V.yBʍVBic$"O -:\gpjslO-XFyD%U=֗b>.FADm})|Ę,vbW Q1Mta+'gg| s)ؼ5b}-q~p 䭵VoDlՕY{l[[5 >-l sRuGaH0*ms# 9Vi Ù)iec 0;ԜMVXw(ר%ʼ]wT!0ChNiFuYNu }0/C9v-02E!Wy_Ula.?r[`Ɇ͍zV6wHcW"},ZaTPYl[:@D.ﳝaz s7Hz})&>l ijt/eQy5`)FYABx˃,y93@>>鱣Ē^Ѧ8:^-E gL["Hn~BYhoї"ʥzެ|P2И3]o45?bk.:Ku( uĺ)- R@{•$16o61. !AІM7, X~jL@=uv!lX}U+`-&NVtSHL@|u~4O%x·Q; DO=>)7s99R\p%΢n^ƃT W5WVyY޼M%Q Z_$i 4fp%r*(>K!Cvn?=Ek,Bf=7SH|C?>5{ TuF1wjŀyZ3REi%:]N^0"=!㨇Ŧ{SBZy$ {Xic R z[0}ԜoΚ#6O^$If z_s96z m_ )L* έNDpڍB5}Ƒ«-#%e&inb>bB} OQA?Ki+Ѓx<&e`>ĀaƻqdgvxNQf=kPH8d*EOKqиY~+l?ң]o1^?CT2T64)LY4 &ދv!QV,DyUyѫvNk=bS[ `:ȱ|y bE(y浜/HqňsV9}VrBwćV˦|B>/GlQEm@)Ю8zgu(14GIx5Y1Y*mPm=q$XP<ȁs0@T57ΩAnKN515N@jUREDtZlw>&x[ =By[Fmz@vuXd(SbǰNQ%i.1P2a$ys-Ay8|2"q:݉V[xr& dϥH9Y(-0FeHIH.w>z}H H T)ܯ7 sk8/(p!q&zzS9F a2WMgfSQ$W :)fiP\ ĞΒDB-0"Jd?嚵oy]P]bۡ1묿~RP$%Pi"#9P("Bi* jqUBHFKimٳ9PL'lSm՘ͼ ˤq%bkK:Sd ciYpԼ)^ɢ깅2MV wݜ3E⛜N!L?I1!rk\-ĥMA3zG Ni9 ep:]d\:NVE˲Hȏ&.i|rD-ҽ3NI%-O%^PS㼠]FELv6uM w[X%+TdBeyfzim P 1 NE.5 3+1`EF[GXGBS1.{;Q݂u.wt˜:I8N=k.@ҮJGԥ<}jdMg8K8".L`q]<'S7^QP :u.OǷ8*ci'F9u91?-e Y2.wPR%ԚP8'@@H12 \#,Z=(Upisr`I+~|*RZ9dN-yZ`7Se4M5e^鶀<4ƣ&+V'7`Op׵NAgr"])@G8r/熮EƗ"X/-k|.y _AR1C]gwϤ\$jBVr2B䧲:9i/d,e449y F5 &Ǖ+HBQĞٻE$XgUpZvk]Mf:wp^Em 9~\Rbުe0q eq bh<5Nvr|G +p!y!;yd-2WIy!s3Om1ZR`A,5g)c!()8.Hm#>m1+".n?wh ތ=k&y>?>°#N۱?J1ĝDl..tIߢ7F fI=Ed&^XT^k1,D*?eQx4wv!AHd)mqfFl40U,-4FX̥k3 Zة^\CWlO?UaowA| ئzl pmqVe/G*Qƾ݆:ݸK^;5Q1M4ϞJT2*A',:Ng?矿XL_֋XhqMnIDI jtNg:鯛:[ͧSs/C E_;aX[-ѴBjǥe%80Kƃ>$5l ߻ 4sM?&M\~Ζ[Q4Et8LaUhm* löbR%̄A Y] 7R͙@V^? 'v cXc͟l*ac`n^_'!0($ 7ZȲdZ(^x5}"Q ZI1N>Qt~{Uן7 $^ X?=6ѸbZ9ȹf D0A Y C YϿOamBf9Rw`Ȳ8Op!!c.峴L&ifJav`<ݿ 31Tt7wk~2\}6;tEZ+`YBboxw;hҜF3al>prn,9h1R J!q8b\R?mnC&o';ʢCf*qNHut 24 a Y 1] (dE2kY]t(}.$xl- X.(֞Q@y V(M~*!8CH e@j),98s?HR }>\8ڤH`c~+#y6$C`w -ybP7~jCd=~4P'V8#+~A*߯Ґ*{UX3Z){ ɒ,kF~{U2uG{TP YMJA`jnLS&Ls8Dr# ɦ i95g)LMf/чn[Dߏ8<ۤ3gx]N 29'k@iR(_3 ̪3JHA'&jU!(,GhڂCiڪ33*1 {6-(v/XS׀mݥD4^kiӜ'$&BpAdvk,~3ϳqg4U^(Q8{1O7MtFp?*;˽J~b#b`Y+"˂D1dl#냫yletE#@nsCBb@- $CX( V^qʠ g0Y֙WQuvCYG6'XۑR aR+س 4nh8}脇ؐZ#!dC{4r`Z^ cu$ݗ#oRQTK,ZTnd]4ȸ 0.M7ҽ|\q=2zۓdgyoAF_J/VWucW9rRE1̺z0zO)q~~.?U3͹V2YFoe]4M!+qs9uNm=$zZ -X0j}[cr]h{F@ }nU/cm4&!#J/N,^d 46@n}s? T> EZC@L_0ULR,+[ͯ_֞vC5tdfI Ӏ0Pd"'.mVSF)%\U}1ǮJSR(e@BdT]Iruo,rԯQVN)Jd2{UH!DX<}Ht Ēo1DJflߚ"`N5vs(5!@eUZv(T~/Fj-Ι=FQ~zz`y"`Uo-V_UriKEA-gyAtydދ[蒖LՊǭcAPlmoyط;2(T k8r7"[̋lyvϖW"sr HpK6 }HM~jŪ$ɺL<7@P֣(]xZ;ΫfAFKV,D%7[W4{Ԍ+" ]5G{43;--AY=ՏB]N?>,iQE& ~da2P_\'jY%|'PZJ AՑ6ʴA=K[ _geA;Ue]%KySmX6c%n\[c?}欚3P"2~Xs38q#m2VmC^ΰ}3 8.ٜsR1Kdz|b`\;G0r AFQP%k1w r6/m)+Tu*(|+1INaku~X#ҪN$pWW]h P`/XН?/[^d@!jtl8;Ș UZgyd> z. zc=,}G!n1nG}|V\)j5c@6DR(&LUEa`n)-Ìzy61~,L|5쥆c J^OJ}X>|ڶAU/d$n}PL Ҏʐ=!ƣG #E/B6;:k9p$29?vFTO1Zq(ICK1%xt*. Yy^xE,f?tc]mNz [G#XEsz1HPG0u6e'1VCcFE)*%i9?^-FW)Y.>4J(Ykֽ*|g5sOr,#>9Og)[ɀFk=S  J UllM.5woj{kkab?k7'd}7|7!̀*%gPuLґ{rqrd6,))qa%{2}cH(1_+[8cbP7C2Z,@fƏ䖝 &JjpNg)Z]AHQ4_ۼ0A1f%1e?V^(`Q`X1 Ѹs CikV% :*;LmN;؜~;6Y>B0ʻMAF_J(J<$#!j={K#R02#ꈧ$^FUVL0 V=Uc>Ⱥk? 8XyĢū VL]cW>/-B|l\3)U^p}01r=nSE7x D7z{hktּSZ?kOˠC} -) 樭҉Zn}'zlz#b 6Y2aC*9\+>^n.Y^ϢpJtΏ>Y9 2$}%La$Iɉ Brm=Vb'=Vo9|%kX?{WƑ !`gg0iC؝x fuf-)[ }_TSfU UU~e=#pu#/}͘=s^ XM]դu72sF)S!PPHlgȊip7Ϳi'0dV5'|v8!i}{WvZw5IX%u{p S ྤ[%x2i[ m_M9w\6CWvqг['LS_e`[֩lc4R<0QHձE|6IzYrM|)҃[ ^W.o@G@s0J:&f\>W3ۧW`Bվk-՜1:puy;$3$}U!z4E#ÚW;z'8DxPJ|d. H@n}`N^VM\sq~4l՟#QaTHGֲ5FN?EPc)yĊBk<(W ~N)LoI{@ #l{˃*#wQId|R05{WWۮvw" Z5n$hY)̵YVun.+a5\dT7)%rH^iOeE!̪*ϛ!,+/;勳.0}~IVl;^^!1]u޵<jR|!R\"j>w;g2Z8J>ތ?Bߍ^ Y]˹oWkvoSz]-|7ׂQ+QPf l%?ûG"c@??Z,ץ,pm,Y5v1l#?ovb%dA3wcan~;ݏ'#6 o?{FF_?etϘ7A0r@m7Pr5z\Vmx6wEZspk 5:g6'aƼ?[MջtCLcY]6k_L3Yr*v ;ECȱF]#m1Y\"<>L;.2(?U鸣Pg1#U"=%s;ϧIٿ|)uD1]-]8["_tj9kZc g]T&=&c3( R?QO뒐J(6oae,ķy eLCQAls ?xiB@yXT`NRF=cT}uA˿>XDhp#ߎzqhQq>e=P)2#t"%YݙVR|<D*((GDbtAkixg>^أXBwIhP>^ 6lQYK,H=VE8`_L>۠5`!o7ժ ISA9D FOmX+BW&T=_?|?Uo;EcD~Xoȯs1e.sV 13F-4"cU9)jG{PJp vrou)$Zg ݪִ^ $!||N!LQR^pv=D}tynPoR)XAhX^voD`{'p>ig.o#'IRW|zy.&i` XCWOڅ"F%m ȟȭ|sg)`~to`8I)DD<#Vvz]WKM6ϻB[JJ1N}7|V.0R}o>$i;nq|?O' &Uϯ_&}׸˒ ʄ(3ZrSI8J{͙D*∞*#P1S3T 15+:O;$H:F&\Q^{jMR}G{PlfmcHDz&PF$FŽN j\Ǜ'FY scs<ϫ16t ཭łdaZm/I^@ DB4MacRY(9(`dn !ڢtF%6/Yt5,ʀ"3‘LHpZa߀Q iƙB. BED5!& R x/(L:^vfz*UߨFUܨ Jh#rFcq@4GFI R &$,X[R1JM꛵u+̷|+̷ʚBZ%r]mQDq8ˍ+tV߫! GdklO8J!Gb+bۋ3b\}"#q_/<x rO/ٻF#W=@?v7yhT2%RRQbY*f̃$+*㋌#uc<<od-˪X<;_|+߻ O⃮?GюG0 v!%;쉷肠 ͗ԚVz\6R: J+G(Dhm|,+w"z|P0Լ^٥]slsxzX^[DJ]m`IC1YdBŤ}O|:(9hPi:ȪG4feTKUvdƤ)' 9ApKJ1696o-;@ę2SjB _T"|er"!"XG 4ֶφQN@izOj@QO#i=-cIJÑkq էƛV *V\˽A$ITϥ-N'qGI'G'dZ'uAG·[ˣBSV[[ 4l1*Die^ J 86_Vo~ ӄFlAUW,ox&L t%r”zq[ϙ@MNaka3"7 zۇBٕj dXJ|bU"akm">kn{> dKt1[hjw.(y?8ƐTZ4gZ<}uJh4Ճ9(xDt@3-u/i/P!,~0_h- `+Y"[m2Nا[XKj \0NtM|s^}=σs vZyQJ[M`ؼLąWN/IP*用 H8}'G(9,2UEzaP igy˄qJ]d^s.DܣٕԳrhm4I/J{' u7gm77öTZ2kb]=t/WџWqvu~X,Ba)nNyxfZԲb36=cjq+-COuz}aJ&'1\̪ڛ?/ 5Zn )_#W4482وz;mY%F'^Ab3V덼P섓Ʀ?@pūVn^lY^Sh-;ekFb;*-YWfi W6FCMUfݑb& -UOFv3 }.N&_A9T̹KuCYQLJVց[6-Wˌ Đ)RF!o潐 KVAcKA3N(C6-Sf9k}ޅYkӖf+<" 7S mFW_d FƂ8jݚ"7x7MG>iEz.?ۇfڼ\ t|B}@{ i # ,\{TtM5]z1';F)^L\w [wa͕l-8*SQ"w|@}nM"^PSms~KVpy/QwX]jÛy&ׇG{,Dc)NaI{p-fH8;=^Tr>YGS0ASMn~o~~ aj_3XΗ ?Ƨ?fهQ=%ϱT kvODzۉq+Y<;vmbb >͏?ώ/8NF淛mKX,SFǓ(ư43 W?[ #Na6j&o<و$*5 1QZHU(<2$/g],ZEeBX8]Vs03`E}ZK7ee U[$UJ^ҡy&tCv 㯹^.voy#wzfkp#})~hLpOCMnCM7`i e"el^ xwDWz7T݁Ci)62kRKCn6# h{ ,~kV,J7Tu iNz8OJ?|۳=qfa_[|:az9Om+DR 'RiYXm ͑v`i(. PHڋPb^I eș8(ì7/l{s wllSl D p.x^Őx W Qa9pR[buEzLCc|PLCM.BXOzh _30\q} +fyP17%0En"2~5e3߰]ߣ E>NoM-!_/4np]+cdR`l@B5FӴpa:;j<3"/o\Nb'-+U8N-S8fLǧK9'ieI%/,g-fTi4`zΦ=_ɢ1LBr u'fqYbIjŶDk}d.dNߵ:g6Pl,xs+3`ȣ#p"΂.Tb&j8~駿}? pUp. ﵒ̓R-)|ޏ>y84,%hDH! o:K,·N~YgR%jL.:g" P&[+m_V'k$޿DsFL"-VH%? epPce]x̰M[c?Zܞc$p23=c*"F<2i1FMX\, XxIAraLƖU)WWKǘ*?GTԁWG[!6H@DډDIܻ >]ObIb!nbBQLv10bBV>|r7^tս?}@,xo)!î`PC%}Eؒ (@ SHdodElM2ad'Z<5ԒFȄY67*:-z^}JذApTӒ5 6d;%d!*GQOt- %_moBhl@*6EFoJij4 ֑9LeoOm}UuxȌ]]!q=3cD5cbҩ3;'Qޥ ?HH=$)AlIwjЛ" Xb#Fd: t5~ӻu8?.~f'ӿ͖56̐?_Ԫl^Z D0\JƓ7!{iKI";@c !HLj ԪI& hK"-I<ր<+VIǠO)@ힰ*? 5Iɟ[gwo %-R"PB*BN=bFPA7r#3 w]t>RѝK莪 {85*O sOРOųR?uR?Z ݹ+]"OS?@(c梢qonӨhV@cfOafWv-\߭ý_frwJ2/G?ͮtz2:8: 켰p<'{*%'v2^>zH&'y6O؄aoM]9 p՚  qZl?-~^_4ۙs_6< *Gw=yB_q?szֶCŗlio%qd*UD6ݴw?)Wo&WZhX* )k|rܐ}D\ՃϚ󔥛^O=P!KO2\9jub{' zOwnIt ѡA}dw[*ꑌZm:߬h^5%*k9ZՇytFthӡ*mP;[9EV\J s[iNF^9#PTMQ;U[yG=gZר]v@U;zzm# hP;Վ[9CVBCQѨ|Xn8|"*4?XK'>v#Y7.)JK|> Et ̜hmtݓKvFDžK]\v v'Ck9ͦO \s\LmioE3|SK+pFKXZvwߏ@П/V<_?}@2LnOٓ/vzW\24{S]_OdOj)'WgɈSbf@S@X;4]]@2z90oϚיУ7!MA!t߼ u~={aVƑUjhOF5͒¶7B ؄DO=k`XRrk&P| hQ(,iF-$Vf[6K5Yn/Oϙ 1&''J0AdK%"ʠ4i;s~Ct/oӫi XW4@^-Mw]4 ]Ų?M x5LhZZlE=iee8 ~j= _0]9AVY9\Eb!MQN򔦍{ 7klĘ p<C}:d:}4$Nc-N $mR"C2H2kʌdM{G?;0K*̀0vy/)Qm IV$YH`ϰQ!ASM"!&z% 1K(VZ U= bhm+۞`X@$5]_m:Wq pwahD(tqy_2hr%3sY&iR4|b_6f y^(I X@hEqB*X ?h@"!}u  (O*H&hx渕hr11fU=- iEpTFR`D'bYƒzj2&nE2n@\RznBoM \3<02#MLAC Xsw(c)`mkF4zXimu.}P7"귴bLThZm1Nj1ٯufN<aRg+(7o3\2& o><,uʒ:]!0Y)&.<ǛAtow BuR1xDQU.fE.DV-@iX=ņNòY*7.FWNldzP;6Aw{2ۦOhIsL=o?ڪ]',IA * \d.IN1+|"U2FRHEZr,cIɠIWE %dxfAlREd}. Dfr[=^KyJ9|Nz4rhIE}Oho/=6&\&,#Ci6n\Lplu8K9 RZgؼnN.d?>![OV;6metE٧ l_=~!,60ac Q(dE"@ ōZlc1 *mYAI"e k߼C!Kɝ¨\ D$=d`5ܙ(jѡ` ["ǚbl>ba&Bx2e$w# D"?ÕUYZ %Lro!FDf!ϛ7 U/EPK =8k-Eۻ6o~$?o$ S.!ڌmcJ<AɖLSMdS"qK[n ػpȼIļ|#%z7zf&2WyӁZGh*xvzz7xn+n;k {y;MgbonN8k銽IN7wooC51%CC߾1Ot|p֜.{ZS_Bx-$G*.# a:P(H9u 9hI#AP̸8-/w ?=98܆62?,$B\6:8u#{N a#Gs5\`>]`e< .y6yZ$+ 2D\Q"Jn7(ySZ<|noK:Yҹ̒et.%G͍E[̾oّg;՟Zp%g9w[8 ';1S*6t_?a꾐:_>}#Bl۪6E8T'KՎ'Kw6G߾ %V8Im eˬ[Fa@GG 暌Gki =D=RcxEM;O.h1nmTo~}Bb%RJQ⢎TfX{-~PGJ-aF3ԍ)` SCЅ6IVSx42q~3?FI{JrB) 8:iT*%QJH <@Kr< ժ(`*ݴ"rG yn88$ˊOK3;1޿=rI)>Mc>ךR˜-5seRGѼӘh`G=9|o/y6'S=tfLOndy'ȕ{j 6l7k)&mɥ3  kк}K&Yl98m=YCfjnR]rTg eٔb9#<ѩb͗2{=_OջK _M$4@I?(LwOCLQ`C.&i;8|?=o/y>p٠ʊ1N+eNO;/ 7& k[-E#2kkܷ~R|Zp؟]l)EգI+tA|J)ؓ 1r^C {[\*+*'F3!Ɗ׊;_edzA{۠7鬺,*Ϲ/Wp >Qpy*ΗCv]́KQ!7|R$xvWhbߒ+Cd!F}I.#}]aN 7d9 V>o>T%y Ɨ6'Pҕ:<q\ÍE n4`vWp+,•:˩/;D0v!&gR>n/:#M_B/Įn#KG0N3sR㊙]Qڽ hw}N2_hܴ\f.3w:c.ϦB}l2}ƯۙL_?~?ݼ߬ 3O3zkݴt9U~K?+.kC HW$@!&N7N[+׹rL9FsGFsi[Iz;NTX5]OmNvminWM}wƨf-ŸM7ܾ*ᒀڟ>߾zLϭgĉ//>5fWZCy7 )Za h"GA:,j&ՐR1D(,pe1Q(#wig"G41-s1F8!}O-ڴsASq]j]uk.T}X8 )f>k1Xz Zhoe,d8/!>&{ o._9 Mn11\cO$ˡ;1̩Jav3sV[tDuA$_~@?X8)bzˠyd0;c|FhIG/RmH6p+n6 1up\BE(K&@*3)j B28i A["06ǧfnl).'ۛy~=>7^7!!;f^wۏ9‚#ԣ>آ]\W -5V )$)"'n92{G},ɢ|XnӜRIM2p9 YaȒ: L. 7}=C-ÄMpŜQ#(̎%- h&^)$s,)j\(3V@psR+SvWGl? =BķU$}gN stZdeaՎqyQՑxdTR=Bf'CI bAQ;ͮI$_F/2!1T&%IʤLȣO,]e8Q;1z\JFlHRQLuԊ.)JFp=)gSP!ƫ]]o[9+A^7!YbU0vy LgNq_EٖʑtĖ ĉM_Uք\Aj4Ήk l+rznK/d*gNR߰lק{j)>2V ph3iw|_ S^]wM㐓g;#_a'"ahR "Uf8%+`bsյ7E.4Ȫ'&v>C?>&d~H@ʗ/A T%$: jW#lU K%-S5 ڱ.h5زVI49Y{{(H>E;;Zꬕ zș4!^K_I\ (;;8Lo A>^a X37|l <|0'VQa^>pFLyg0f/~}p4sjX#GeXo/mt8[^,csgmͭ~j(3i ߺ3EGQ0;N h'7gm"%;]ͮb v_ϭZff c6kAT2d_Pƨ2Am50SEBMCXn3 g뛆P~/a{+&죴?QTާ eR$0.)OiAOll"w!c6=s8X a͵ 9Б<_t"'=K_2۰WZQ 93WRmR. .LrpeUBE*},B(jClʐb&>GKC ICJkiNVKX T2Ȝ˦!g6?nɷW及,Bҝ`r9j{1 łYЂ-=b!QQQ'L (lt:< EgxNw7Hr> mP*OnP! e-șg `3RHMYvsLj ny-n/P؀R6~G$ Qѹ?xKY!Ȧ)6j+ j]2)Y ExHK wor1) mu;u=a]wYsAq]]sYDrAҭ)*`݋91%[\ {Xe.ꄴW rrUGo` 2<,G?È+;>Ȝ`zF[#vkƾԍT>Kt8e^9o(iD*5Kr ؂S^qOȈ}J̠o\^Y !8di['u{U󟗓,7=7;̓Lg\` 7by&rN!@n>)}AyͯWxh[ ~f þɆqr!.<%nMC4gCA9=M0^>wJ4vCx';q@)9" g{y<(I) r@ 3giv "<}Q]Ow1t=l<̀I|,3ŸQ/Ѱ'Ͷ>6i.'1LҨum,+'$ cJs;YMDqi<#q`=#n^mV50Hci4}JWvYsF%|G˕hntUw0:9n?D{Su^2_'B 4av>)Bgww$W#6?z5: xWJ?O#~o/.~C)Z3sl:{Dܫ?-eIA& =kT hĂ2]{j[-rKp%@=;( n6qP8c߀f؉wã'EG3: w>1#+@;]~@\N`n@yF{·~B;kl9m0 XhA zߙ_h7݌}c&jS#OzV6z'N_(f8*^b5c/a! B P[4䜷9WmԟXvє2se) 4`CQnڨ=e#VgJm8VHF/_U2kI[8WM5im5c .fSFEsK0!_j_OUY"{8[hL0M7Vn[~y--"NW]FoWya3[1B G :띦sugB|Jӫ[qIqÄNщ7NW_rެ&uWR.+I8 kuynK F'UALcN½ʻznmqaRF2%~ DIv"ǽX'N BT[]=g/)#lm17bM9.;8&$)j&+>[)>QRڡ&*X( !Y.pB{ qd9lt>XRY4z2d\Xշ8QLRj[JS--р͞Cτ\b-=z%zx@+,fUQ86M eAμh 9iMCX\3c%SͮGMI/9[}/ppZ3~*ޑ[A+uk#(e)8}"p 9ݞEpr0\WW6:gU? MH ͊u% ZjYG~qnU>ӻ/x3p=K-oL)BXޖ۶ynR/&zlLDc&}gu!?&}KQ2WY˥9DY1;%ڢhqGCiF>!Rz}<!kS"(=HyhuMUR*~@l̑J53<  h7gߜ}1}dY_'Y0=`}0#GN{J(#:۸#ptУ;L-4GG__.Ϝ h޾ء.#bxJNYz婶m7dCp{@;L ` f7[1|n@y& hȡI2bFɜs=Oq@v>,6ݏves?yaT*\xp 2(eɀv2'.m+u4kDHh@v>"7_*Y*Gz_JD.7jlJ|P*uXAzL}~/ď/߾N/ʋ÷WE7$FCAhٯ-UX }/ <@3-|Xs]|.(1K`S%9hƀ%_+1 ۝AsɅjͦ1V抮D [1@DLllkΔ"!oIѶjU ';9ULYT&c,TSbA(T qr (|R4[CI\ fTϫrRpUd$֪")zB|\+UfęTKb}3iHanXq摄89q9#?'b 9tĺل^98'YRQ̃g7A_~?f5 Wh~-w?{]bJo & Ϻ-Vw'a9,@1g. :%aB!atHMs,s f8^ӔxG|磐ٱw4vhc3.Z(h$qhA>zzLI @T*L_uj{T!ep(bksfAINW~!* X n<\lC*eS)`bU~+&Y' ,^Hu:oOWHmQw'/UʘgQ،Yg5G0>rz65ǯfof.߾?Z}urA1ϸ{Øma\̼I3Дg٤"9ͩoR @ d@1V -JEFKgz9ӄJJ劲tWJ,YeL:t,w.@+^e\e?O -X]CjsfMkP%*&#fodoaK d8D[.g˰MoSM&N]?ț[y=`\7D<`ߎ#NT=)=qF!bM;Q61H||3hF+76ۗ_'$oѾA gٖ!jĬ) hnXjۢ&\hP+]le@hˀ~_/A/.B;N (!Wc)& (PZ !uVudv +}uHI$Fxr[Ԍ54lwWDqb/f'R(0s˜$_5xo]V3zbC nJda}S\kP~fӽcã Nj¤}D ϑ_ 9C*g/WDm8#C 43x- ^{PRDe=\AD@,^9НS KP.2H|?c)Җdfs|s>-Olk<;3NaNax턹=lRp!+hKy#ީ0,iFS|I `O՗A$څǛ|o?U خo혛E(T/^B%= ̶WT"Lux p"'޷b`|i?ioښXJNǺ%]uIAD()w;fTO׸3ޝTJN S NOC62w-+ɺAwzD/>R5zq2=&ñOg}}mBU.޶sTGee9E|R 'kN\XY{a%< nЏq2~Jƒw!TIjZrAǫbsKj%SAiZ]%` 5Qs%$3 :GN[-/J<2뵓;d,49@''i%w4#ȶ@gF9*(:Y$!tɂ %x|"o:ߚ$OQNT @;fâ݃Mx&_j/͊tḑo@<Kzz!,tIMXN?.-Yi!0͵Ɯ2;F nt5rH 4~3˳v/-Ͽfg55!S;I{uVEF;$.LhfPϾmWRN(^/nGOux4:gTP#4JτRaTeI,aIBeX_gl{Zl8=>G`T >g- x5e5-XU_/>BRB:I ŭoR#B:RGh%NG|"^kNɠE @صg]ǛE0X K`bu}c g$7́a s`o uۘ!<Έ6&+ąq(% ).$J sSv6ffde#U@B2?#@i[{,D'¶d򿢊':":3)PoNG0}Y v}\on ãI/py2F~8?_|J(2cltq(ͻǾm`N]ZXD OL3Mb9̩O./~I\fo<&R <Ӝ*Xmu_mIuX$ay5A CE'1J@fVDfly3P 5ࡹ:OJ|0g7#sRxvfUe/O,MVx欴GQX|r@j rA-#G8hj<gK|KtLIaS OEb!rQ0)ydzHTbTFhI')!Ӄ\+V)mnk$k::_jMֺ\vq'ADιkRÈ16T3c 7yo`XdVwѫeYJ-e]IybW![NuT:FGۑKop:[FQyd -^؇kX<)ڳts*j$|Oer%9k-| CP  ]av]ND$ *`kaQV_?DrTGlKr @P( >O~i4NqdNUfz+OON(9fT<&zt:@+|p$A2sǛl<-%Lr/-x)XΝ1;ĠC~n g0pnQҼB!PCr6þ?LP?hV/́H%6tyKqɩP;S0\#HDu "128Vj=SwKkSfV!8 +A4cJmlT8+v _Í;V^'kLJ9tjuApFfҹ8NjqG>6tabL+"vS{aU̳mJx?z:VonM5| xgaXBIaȺj>YT۞Vr25$cI}77|U:oJ:K:+qN^G5{v> ?)'8%+n 2.?+lD#çOjҒZJlt'" an,ho4q uf3iKî8-X8"( ATny 8Rb RI2ͷ.D PtbtRL-յ[Gcq! -<۬E[3 wq0G܆96FܺpFZs[,44EN%ESA26%JxnBUpk_;44NaNav;pQK'f%CkK81ioBBJ_rOXI0J,qQPwT.L-v8kC`+,:rAYo 8(&H?^iE0tm]Nԧ]E]M?kH(E,RbJbʤ4wc܋(`E;嫨lFcjzJÒd_ ku򜆂i5z2'jD a5Ήz4l1*] 0IKE Qf%y\D\]W.T+_Qf XUiz6`\SZI\Q@ ZH=@bhP\3O,jwms{"rӴ8@IЦho(%EcA$k%[6wWg Ѯ]p8l~BTa`0SrB8Zh6!ĥ5!G!4`ڬMgJQ!io'S4kz2m^"aLsr`D4HM\Ϝ40U$V@a'nߦ6P'!V{1%6Nm7n$aWgkq5ڸQE*mۍGlK&Aep|ދEq1N"Iy_yL6)ݙ)_(~ؔQ4Qua2z? F}AbGr+`&}`OGZo8JP!B,>maWM)x#7E5=)ұk+L#V VkHu{9ƀ?gdxkO_~|TR;`./f c]M]:WO~R- OXC:ϣA;x ,kbsl k8|˻Wz3Ư}_38 dFt2kfQn[ܽȮ\wN'1}V%r0@y#xp#2  @֛f;!N ""Epȇϒ*H3#Zհ,M2'W\G\p>l0K?Gi9pX4n]*`RN3&٢$j'\&b' MwG&&LᏂ@̆^o$r0!2< Jr#itU$s\(\'gg54xsLgσjqqr@#&9D*&4ʆt&z⯵%R{koî}ld+duS@(QmIUAߦRLµp6@gÝX*,7oM#%_(Y8[iQ g=i:rU)<|.ˀrv V"(z=oTM ?rObR;5c!x"O)S} yf܂d4etdV`)ịjݵܬR{9xyEc>92K:}qFJ)O1)~{AV,kkJJ**Q*$t'"HB09KUMR!T ^KM~sx8M%m2aW?T#X5=(fjcc_yB1$"߳'e"hVzx[%1pd뻠>347!VroO 'Z"in<+I)iȞ(XRHE_Rm~믿- q~$;=sSb[/up6KZf_lﱵGE郋j*6]RJ=/%T&RƱ6+VBB0⽂CUcpJ`0o",?WYcXcBq)n)X&`)eFj+3{_Rxe6Rjm yף*[Nƭ8 DwOxmp94.;!K`S%;|εWxlW9A1^zC=aWf4ĕk- b%KG?FOxfseW R,όɵLqo))wEľku<x;>0}cx c\LAw4RːKx]]gF;&TN.^ENmV A۫ݨ`? 3`̅lǹ56 @o t70\l(eVhB6O.7 ؟.Q2 o%PAоϯeRʾJ DkV~[TS^(/KlmFfҘV*iL|LKcZQfT|R{UӘVv;OcZI~b)kgJYU ϵ-J$Soۼ(Q}'D #%i&?M>_fBKe.ޟMIt`Ś-v&OW^ĭX PebBƓk7˼QrlIp&D&➒Z\+\ dNkܲ3n6Ѭii3͘&S c"{D/tE! ( U-6ށ n%Chr&L-Q?D{8y{q;  W)Yc.UU|0Wmߊ+񪯄5ve ]ˇ>L\st* MVOg hs-bY>6B9"!ZIV:,}}b]A b? 夌ЅbHw@_" ]oRP|hĸb=*m3|@%zG՟8|ۣi,s@wD~ZlvVg ~yxoQf3_m }s;z5̠>o`7|=z>w.a[~-|\L7K E}-b/ث[/&z&gaC1x.7WA}=\ X(q336v(W_zo~z[hūIm?<|9[u>|:#?h~?oѬBK_VXx:.^|`?U<|\\!O0JiNh7:ٚb<~:9M>,TƁ0>s#?͏X^| 1trX >?kSt3|dp-V\Ѳue3wn0ckӓԸ,xdP(a+f+eu:XCGqp8 RP/LKQL]CC<8@}6r\Jpo7O6b_DBc͓Xjڑ.](*h} L2zoρ}^Q[d%8GIҖIL1z}$[&2ɭLAfPIEtL,X3&u,#;23IxK%o[g%SeC6Ɇ.#  DNJ=Dd"[Y&˖E,ew#W,`JN?^n#Gg,$یxf}K1Ҫ۽Sމw lH&d"cL92 Rs],V Y1uMbJH3aN3^u帔``'vWv芐s\VVکNkd=#v#QO'oE|wǨWX/oU UZkC:˾R1bDTFu©\zJQLgLkdmƍ*U<97ΜMyz3g`" ,;cFVH0Ў$.;'= "P=S?5aT1Lp}.f&Xn|}s:wE 8^yEU`].:E}xQsl +7o^K6D!TAo &N!WAk*R9 K&&@XN@Dsd\Y)uELOӕbvtF,ֈX/k#ۈ6bro Y^X{yPl- ^潧Ao!d Fz6dqGPR}i J8)-lYdUIAe-{ڽ;}ڶ{k2Ci7z#6cZZ9e/I9yld0HBTjc{}"omi"[Xge-lYv_5trr!\Lz3FȬ>2=nDHJ*yK$oEnSLїQrD$id%[.YKO2-ld5.9!g-\Ì6s7d $ K‘E>umm*T.Y(ϊK˦lDb2 HJ=@Sb o6/#l!0 HB1+-QyKܩZi%DkQ.E /14f8JOCR$̆\b@3' nX>qSZrA-K.q^+u.}J8eCΪ|E8bۭ]lW*şvVE Xv_Ȃ<Y:$|tT C/&>>OPo A?ATse'䌗7?8w;C1DDphwGrR^]} #;p5/ub 8$R:ñVaSUJS,=yׯHV|oO&]U?si?sƍB[z>Gzh+6zɧwf0dhڬed;5M"mi6L3Mó m$RbGn@.2d 3R(w|vu_izǷGwHOv G|r^c;1Su4g<"pe[sY[CA`Y([>QPy\rNџ-H|h'QMo?NDD HqpJbCm. $(PW]F9dik6UDc8Pӳy 8߂ aRF߈@SV=;I+Z(oGӏl x$h&X &]9%'ⱺyQk?E?}?\\M~tjNo~u&&PuL/h&C9Y]jLeQ2=ʅ?ko^ONG}*ٿK?O7֚WS .MV \ {9II9{Sbm~KQu%'ŠX9]]wy*Suo~jrK>>+"|*<][My$cw廋o|ҏ%0c?~en,IΖ~Dl%_?4qΕޕSW]Mn[]K{]6YwO46Bkj!픒tTypS˴h$$qoFm7UI.qb ".*R maD,Z9ɇ&NŒhJ*‘Q*^e+{FR-+ Wc` a |aYr.I2\ʙP0Ruѧ-UmDtQJkz^N0ZFT&ma5<)ѷB$( HgBI24Ct?i)JgDo14EGuɝ{$I_%>(O'X.YdO]3;ig-ژ Q"8ɀwApg&~ %DI^keM],kt6 vY91Vz7-0DK](`tJr5ޞS~ӷf~d(N IWFSpuZC$~3kgV&8gJ ,d,sAH %FYؽ&F Fiؠiy i6&5T iYvvZ\ijvz3r R hM3y q00$̷FOΒ0,7]mSNΤBr* ڦwD"^\t3X)vh&@K :Dv3d *b)S3%Inlk- چWJL /9prIϘHggș!B,`uI n#~"7HJ°Rm wBj)wgJY,[ 'L.1.p(B[*opgBD ]@JZ"De({?Yv<9?϶PgFC"m0^N`EmG6y z/ uIqjC+ B&U|q e(xiu-fHkA[%)h^˲eUXxV!ɟ#[@:PB:f?Z[.X 2C̪ф6 ?#<)b]hPff%dv,FkLIF&TTV>})5K76JPK<'s7 `,-N +K=a  1⚕N^Q d0A l+wS3c]+w"SČiZ-Q]]<]x]]qZ4c4ji$_RUE%<]i8!'jeQDM(*L # erGL6ld^b+J fyۑ`gp$ iަ ~$XSlf4AK>9~Jw[xAt$5t@ S[bߡ42?` )6i".yAh^)`2v ׌^W/v,ƪ}c/O?4luCiNtՏӟ41>}5ٌY_L}|qJ)jv-濎ywv?,@p>@( \ Wj'<)+hׁ)ԩ" -`4<:jjZ 2իhnVDDd9R]@3i,(o O/*VӆVH\m; ځ-l!zRJuYўLLڷzP77p28Njd*]WLyR ^fpҁ-Vbff!{%-Mm`Ӈ ϩ$ɕ=?\۟|68)ĭa $<.1I{\ ڼǻBg(Y3fF < ~ZknŁ8SQAY),YB[54ۧ>_Tǁ8gbk&p7OAYF?-ylpn,K~YrՇ=ȹGb3/q9ڎMN6GWOE8GVAyؿ /w@j,bGWπ֌m]nVB8#LL)uRexQ}WTyӽ#Շn (ѳ;R{, ' 23&w 51(8N{1 Ysu6iCiOײbywɊ%9o 9sVx'<׃Y03\fHd5?kW>Ǭ`1/9ӣssLB!;1G1OHz8Ey&=N> N}qʱ`@/ūV 2 DnW>d[ TγR9Jk1d`2았аMlhӽ⧻꾫Տj6؟1kN1"J))#Xx|P!zǴzE,*x_@vЎ_B @^Τ ţ=eQp&B e̓EiI,&c]yJ@l {z ҿgO$qQ7ݏG<iN݈[0;F*u mErx (xr\ BPL#RqƝNrגt y&Fud՝\O5 ]kkr1DS$ @SR4xᄃRiWʒ Iku޻ZHG5@8 l!ؘd;|Dc+wVvCS;ohGu+Sm@+}՝gOuc~ARZR@b1h I_$EɵHTz@osrLS4Ԯ6 Rmn>L팧9;7qˏCu0py~n~-|=;'Xe{{W_-s\E[ɱ7x"{^|^uKyf{6!|qF,wJhnRG!{Oc=[?ߖw?I|xxwqo.7sO?N}1u2 (\'p]7 K'n]r8>\\M~#+}yo!@ /q6zfz$Ip o !EsfhEU*ǮoYݸܲQ2Al&'pGf=ѣu&]PFfdWK'_'+siMQrLTQٿ\".\"w粋]<er 1@\Ye{?)9^|!6yGYxN9Y-+W,.[\h1s[U¹rխxxrJ &I쬚zJ O#N,J4۳U]WOGDW.I ld)yT(gPق'jGjY"rK2IKgm)$Ç|aK '6M9hpU$"O%N.`rT h B3 cc pz0pq>^؆8^"E1SW[yBH6)Ȝb~vߓ3Jwۄ  G ^,b0' *ɔV| j?S$7 ئ wAx8;,Hy>ڞݺ 3iF8;=LH @nXڇVB{(Qo#jl$G_ؾ+K^^](:"è }==PH3iϧ[e&~xd P4gHݳ4e>Տ]:~Eɔȷi!ȷ8Wt!-G\]QjA8j$Qac,pZ #- HHݧ!DQRcL{1;2 æy1`.DjRg4ZGFED:b<4ܚDha e8q,ÆkiLN{oZ4C.LuQjkh_"CB"ZK'̺)2:ɃBowκ\ƣlhL'3LJŧ_+#F>2XOo/ïgOH* @9pq3ARNT~޽yU:Xbd0AodβSB|FYw p{nQRG|4^:30<b-; |Dd@@ڵ|ל")OF<a"jkl%C.u,|%2ybqor_Zթ`U!WəY /,֒hXg<_q,d6}L&ՅA:>jXn>hެ^}DPB'I3'D|' Jq%6}ݮ,TdO3+V/f]G,_bn6zZ~^գ/{X?( R&Bs#c o~1HS E0c]'L\UYfU+.p *@Gsw޹V>`h,nc}_Lq\ %ɕ\],-p{*q;ҁR5\}*<}k`!XԅyEƒ( OW0KҖw\;P! t32ŷpXɊ?-c VFLf¯LSUXk6x&SH&z ⹵}*8z8^4@Dاx9?U\bmutv|?[{IXMnB0tA(Rg[7vdO1ԣGkZC#5 u"& 5‚`uPB+1ZkrŋoniԪh}ah5<:Ibn֣-j0#VIlj@/j$z;S!N> qS}QVZJ!9pY0AM aE\i ƅŠX7̙ !824Eb e SJjwg՛7ۥ_@s`w:ⳛ_0i0g˿eFQiro~~ z8º45a! +í SH:7>|+l VE 3XBxqB5IHBJzIXĉH؂|y-n˳v5< id4: a&QZ+9Ś($qI)ӡ `=Ym+G}q< NjceX[SB:6Za]&FkDRYh?|r u66~c!Jw•IuD1i"a>YC4TL*(pWsr^fhNQ\ ) $DT, €WTH݉:?'J::oT2ƣ2b0~/xY[aiAܹ%+{dˎY8x{ |>|?iP }kX [7@UPJt BP\(;glːXEF%U @Z;D>i L;gA 'k 0FOr{2ǂrv-)B :'̣-Hn{Ě} #DȞUכKz*Kbg덌Zd^?J`cD\^)naPlc`_orNٯ7RJT*n!3(b!SRDPkg9LN=&u1PabKP3UO4}h'UE뽶h"@٠Vw8kOf6M>l:Y;Bz8[|=| ̳}9(s% jY"=|}&Edeφ|R|cs;CجhܬnlHt#*uhEdB* T[DhRL҆$)UĸRDuޖk~Y rYyJ5˚Hw߂0hn/ `MY R}}›Ւz0#52_75_͇Liւ.I),{ ppkD5[)NJ:vjhO|'l/7q5|J4 ƨٔ[<לDg|T]cDƜo(UZP'X!DGqjax1A( 1RHEv1rItj92Idڈ$F4cqqH*y(N|a XyOcPR3+=*ĴSejK aF'ULʲA26K]WZJ%S?@2>_i-ro0{oZZ&2R]xؾlꩮ8Z7a顪7>bfE?pae攪0F:Z#Tlt#M z>Jk՗KҚm_).mS*;mt*n0nޡz+kOIFEiAi}dnȽmYnqW9{!Z`ô# @D8N;6HWcbCE|l֧lw0/ryi{U@>y0V_@g.cE`{2j GB[~ڭ*${e,15sw\XՆ*#ͰXc0j ॹH0R) l(x`9鳲r9U , ܗΫ]-ɐϻ4x?#@,8QFUBK0"J¡Uj0v}%KtG~,ysqaUE#S-!wZP*nT;ۀqw5bJTv7wZ ;b|KbDcP2H $٨ #&kPM pg'ZZV!bB]nVcݧgl'(ȴv#g"ۺn0gB" '(K<8z_gN]~2s]/}$+Z2$+֍k^<·t3%ަk:|GP&Dg!qJM.1vM?+S"%Rr!t{>X4bKpu $@ϪGMl#zΣp%+ˋ{0mwWG)I?_܍@p²\`0. e]Hq_dA_R|vZ7`k(gΧQ<}y};vԎ_$ (E=Rv1DOn<$6u<^~vny2_.܏t A]jmy)p/soc^> }]0%X%LwoC q2,>Fwk} tˀ7S'nbWL'|&7V2vr=줸[;;bbh )EgJUlsFkZZe4ZԭmCK5<.cDr/7N.nN}u>kkL^LDH0rs9̹4aJ}e\ٕcƎZ/&W~A5;1E*6r~ݡ#9HFa0*(mH1[YxЌ,{n]Yb SZ>ɔkI7/gZ6jQpN߆}~dsxb,@s 5>_2ù"}WOn;Nr!:aXcX))"_6Io؉)S*d12B&ґODkP,8Th$7X0k̓ D0mHTP,ql ޵pCu\& &G0|`b>_*űI`aRs+ۇ ׳㥶}kϖ?/w:h KU]q)V=VkF鉡_H mVm\wG#{vk[Sݩ&p#]cic114B8 0Jb%DӘ҈cE:N]ui R@@@0 M_Ds2`CBduH@"V>̥/ѠJZj㐅ļ2C) J̆\3&DHG#քoa$HJ) )9V*%` 6J0De58ZJ@k__1ܟj{f@!iRAH<q4h*ad(bX("alR(F>_p5T3PMቆrq WbɭE<}Z>|7vn&My~ݮfF8Vl!Y() \ބۙI"d]ϫdOm|8(L"DBԳ3fS* 9ݚOO(K %QO17՚^{|~ۡ8nNƣx95b.pٕԿ/oݸ%3;0z9.4 Kzfln>1nB㺽9om:鏳-0HÏҰ»aL6 FŽա5z.كI@5wS8 L 0ݓ+KXRk*+YnsR&)K`V7rG n]n؏n^1emBActnta (62$' 'aϣTSq2+0 &namc}v &.BX=ãS  F"2w}0ʎ}E}! &NdL# ]osÆ)tdv];K"V>I:[b '2%˞SJN6&ym6vn|g2CuB} ӂS-,;= y ZXY* n?t{W?\f:z9=Fi$ XIo~zneKJԡUŢ,˗G%3d@{4A\T5U%`)٦}#rYşk0)FRa!/x$p/\uV4IHc"_o暠H>yrzCxGŸ3̺^WگR>nŮk-ȵ-n?[rDžL̕0d0khA(Քְu902AYk79qk$KB!MϤ>  Č_tMv|gYbHO58>l=;;)x5Z^Nho|" NɅLHԙ3Mg-,UBc"%;ppͷ~}ä 9?.ϱa5H'\f|0y !2,y(]ĩ (HVBB9ʿ~> U֟f */踕. O~e\eO&7]ͼg+EZ No7Pylԅ^xrV@L6c Ѿm/r0~IC>k(_h27LR$AqL/u] ` Ε"= S@jkoVjkbɖܶUIUƾ'M'OHlҖvS Z @_7M egsfI&Xj53iϴ0Rƞ_8h_8ĥ/!'D@9˶R a>=mpYR:?N^RC Ďw& a =涀c(!HPi9&jD0!yƫMw8~ABF9XZ@-("ˣ(Ҕwǁ DD~$lm7j6y2vyy F%4yeRyY*D Ĕ Basҳ*_,뎎vJgz:m=`q+H`ڥ"T`"Tԡfr^r=\=w2V҅vhx-~HcW )8\G_A)!25+- {Q>+6ޫ\OZl)2~RaW'S\ ˏ.ӗ*I#3#;f`v`q@"e#8#zF É\uc[p(bj4>f0ܔt6b @"̷Q=5Ғ x˧յ5&0^c; Aѯݑ!NS/O={>;hU-|rqS C9P$\ fljH`16~!)YM=$ҜzvMCr _GL,w!\5K Sf.;-6CAK'@N0q"e2xWo3(3Gt^eWzXNLcA;F@h s/xMD#ұӿi IrkgP1fΝ QNB,:>Ҕy@5OjqC΀qJ>SɇDP29o *F]欖2ICLwr17ݮ7hP\8""FPPAmuGzb$&5:1 @^լmtgS[@ TsS Ɣ.Sqkmͮ{zKˀlV ( =j?e5JƳjJ71_ >W@+*όYiu7NSHМ&Bf@ yFhre2>of޼N-[2 B$14- a JQ@jK^BUsڝhhYX9ewR|.;:4;$Ar' ~rЕ5̦Iw8 {'J\@fHUHX-)mJ ~<_Zx?WĈNR8ǀC2~Sf<@JPachL ᝨsJb""](q|8R| 0 bQI}ߍPpϑ:zTZ}gSG\ _;g h[{x!Ӽ enc!XS9Vc>~J(D*IMEduPY4Pδ\pHC$e9G"֎ΪDB3| ;oÑ .M-D"dYa`s9 aue O"$e T Up5Ϻ Dϵdp٬A,pbkv4C Eۏ:y@@VDPMsH ] Xmf ȉmz gص 'VS/f :G@B)Emcϙi)j8KL9oQ@BN;WWbb$E?iӏiw1+WZN׊҅rB+}ynWI!C"ڜE(ju‡j%=$쇚U+xVK!*N㩆{k ;#h)zWrsmgq8S<1X3?ڕ^n 벂0sL.U˹݁gT:۵;Ю)K߉;N.Ž8xfD!40~g;L0X>c8_V3jgdzRLDr׶OV?5Mp=,\݂֖#O ܃e2S/@΃4?jsy4xkʬA2lr3V" ^i~-ޚmRXVC/rhlja1l>{Gy JwޤGU_!W6YSҡ?f_}p4xUcTOqP/dWz[3 s_ {5mQȟ݊h }""V9dL[O{_$#gաȮXv a`Z|)cU0ѨR=?_3A&Jhق,o-?W?ϳ9܋V+n5O7po.Dsvxq5kŁͤ-m`}4(I675(;M烗 YTg5d|ev`fm74xscɯwW>FAyy6ϗ#_^Co]i[ST7w`ݷzԧUV1r!yVuX;7VSbe5 ' /2efMh ijչl|>DTZ]n+>͍9^nJ )8}" P7 #)pT>"&Vk2\/Xv#-i ~~7]15j1+R>UL،I0e %, #e{k/6'0%51DJ?L.ƬMt*O\Cz &5kAV^@}(aT dXO !fg}qmO>[~>gxWí U( |#J!c@e|xMY._)-A-h1qn&JS8?XMgnlƷI#ʋs܃b7g˓DPԲ}mFĶ[/1 ݃!ia[l~C}w` Hžyhh.TFcL_&䎘CU1 =!A.5f'E*nrRHQJS @R2yƒ9,Q +'t'z>6[>ng)՛q"菛m>X3Ǜ;ӟ?",[eh g,OhAΙJ$ 2Tyc%D+HB舶t Vw(@+n9w+$9c$/]Ufպ͢Z8-nb-apihul_m?Ŷje6|c\vQ$_fra8'j -Dٞɻy5ƱLxग{QNd7Csݥ։{\:)zF &uSǂ{x)0̀憰 7 a.BS.LKpe,\WN_Ϯw/äfT3+jpsָw3+K-\o׹ޝƭ6JD>]TlK̪pD`J9$ H#{AIDth >Vh G(*Z[NX{_<8X͹ohΛ۬h[t]hM ?Nu㐀-^U1.~(Zܼ^ VqH "x͂ØoiCBO-ֽnS5 Zm8hzlsz\k8r8 {жӸ LM :rk0'!mq1 >Xu?l cʄ5CqPV\:-As֬~;9fW1ņe9QG@ q HBzGcʴ(m_%!1n +q v (㱃q=eƥ򺞐zHρpՒ 5*PPY:5{50>['DJ$ 2G,뤜|R+}ϥ˱t]Iۊ?Nl6 ]XKLg0y\MWn*O6P P匥]2$T"ъ)9J PanR+N&ோɥ#*8X& ]L6|37nZ]]ʮiu0OfsTXlp6]5?6#dmm*3b,1$}J.Ynf%?LuF ұlR[i˳ i˳zI XuVdFq%parvV8.\ C %BYXvM7m'+"ÙmgBGb!%`D` VvLQ4n5(9Q'azIJ5X(пHaB[9ɕ[dEtwN1B95#{e%ӓ~Z,e_Y ڷxK\SHg+~w217w/@,0)W;&ÙAm,qog3<¤wEo;3\bۍFu?]mX=yw 9Ă ng;ԸN`&n``MFk'P!1*B(dvwQ=;)&Qܨl $J[MZܮ Igi>3WѰXX~_ϙr~n\>-/cnH/Lfrij, (@VHm-U|]:׭ϴ0$^MAo&i+0` ݦ/h/^¿)dwj{B΁5u$! >r`RyV恝/ yqA*U֔ !j7 `y*-zP\b]dKx׃[q'06[_~{gtgle;l.tB gO/if ai}>GVaD=5 ZSjq8$[4mCۨSOK>Z3g-hsœ8 5 Bu1p> 0P;iUO$2cL& u=Q+}:+E + .R<$gl7~Uo':+h`. QkuY"@SONω~HkI%Cb˯yJy.I; )_CMIJ-N$Q4͓`ebBqۮKg| `b7K핤K#w Ċ6$%tH/GO1-y!4v*:J~c8 )`%!T |M" |:0c_]U #a2kcx7 Xdé0H {0se@(1Cs i\c<\ä fث ,JB eʓK `-to.kr|YUٚ\~<˨s$F\kҌrɌk hSNbL ͥPrh[raW{*b mOn4έpOT"7VG,gH-$$]`EISB20,UE^)_cKi$bLEB0MLرkV1M>>4I4Y@ ;#%EbVw lwJz<PֿQfu7S.X3T,Ye@7V{1WBX1{{ss?e@`²ĤB&I"eP<^wjH+#&1jL^CaXyαY0V R8sd-sF:T{?=3(fD۹;!D%v爂@,AH w퉑)$e,5RSM3lڒQRBwcQHT݉Tre*%&,q4P1;x&9M8YB 2,Aې,)3CMTX:9L@ÌBX%L 2SLTn0Η+׊kwFhz@x;0ٵU:vj5"|1HN.LdW'WbC&6ݳ$(ImjPYtC\2-r6`6*P"5DwRIL3D8˕ D-֖'D`..dA8'c cY,< 5ce)Mca(bPFP6A *Yﻥ)ʆ ['&$)yИ!;&m<ƛO7޺]D͆ܘb^@>Ii#~\-ZjYnta(Rlf)L@ $fIZd9t::2%i R]M':fZ6A27w/I6]uč ѩ1P3Gie{3Gk4Gథ6F6ɺlWTqCO"#h#n:͊%wYk=I; \{a5[FO_}Xa] ZE'm+{F9$Ks;aJ$R!L=r%Nyiqt?_.W6]X8]>7v0Zyۆc|N| ;^(1k4 -_Wb %>^̰q&h&qN@ ƵgmA5.qE0u#8PɶƓAA E5{Fmٍ;<uGN$p׿D_Ƹ:aLc Alkt :1 6@4f0P)@Dru @R}6KlKB$a+I*  Q~HMr&h!D rCWy>h5;Mf7䧫񐰦Kj5cqɖ&1jpIqVWfnk` h֓KkW;<]n|fvn:_B'74~S{dZJF+ӮDӝx!B}E4b Q4J /B!DHBUS520NQ1vWv{О'0%m<F>(F~FV !Ŝ^j>sQeM^ ̚faccwP#jh33kB=dG6+[h+q,L +ŏVHr:_hS!u'sٰ&6zx0)elxt(F0<;Sمޜ!>zs%88*sG8o\N|f:7MkwM!9OOIrU _Rv_ѐu,4 e ;s'+-#hs XvErD .i-al%hbmMjwd^.zk=EC*1Nm:]DŽQ8" $bW:ݘjU9DWN[H*T:ƊBPZՓ")@8Nb|]V|Jl(P u?w:( ިm06sCHiǐ Oj'U?_ܙ 0) w/gvhIu1\,ɧl:7q& ]Lޮ~9!սlbִmU9Y]RֱLƛ5tGpFS]p`b/&?}wQ꺕fλ_<-}B EYAR9TVe?{֍ oJ_\Jmf7m1_W10g&( 4X{b5mb(>O cZ즰M:`Ĉ lKd@USVy1Ճ'\6Z \ڬe>3M``u:.σ2] M PzOφ[7?GM(c>DЛr&zLs֓xʽxمO:u2өr LU_n@ʹfq\zC~ç0mr2:r"Z$S$iwJ11h6FWڭ\VCB\DdJ8׵J11h>wY6kVLhv!!G.E2Evܙg>knۏQ0_-ej,Ȇq{A `w'DB 4xT+Ӳ!;řNF 9r-)%ճ3!Ļ1ɻLB1/K hGK V$EHngqbǽokZs/B}T'NsVJ&BUDƎXcJ#a͙`SCJy$h(e¬UI `).0Ei`,LLA<锟`JROf @c$ْhVm+D`AJbuS8FSyn7nrO%+HȑhL rd]iAb#:cnya޴[1ڭCqi7LA\b#:cn"O2ʊ n5$EDh#v+1Żdjb>]yk)2:#rkQ35FxcD{CLl"-ݷWzw;:wxp*>I5_φե=] }"NY 4Nc6$DWk缡g?*oA&zuϫ5sPxv<Ո&@YP`< A1PVB(pɬ%UYWk,yCQIw $: )!/ Ǜ*ߣ< W^$!!G.%21?xyMa|GQNm[!i#kVLhv!!G.E2ڍ5=jR rDt6ڭ-ZjHȑhLvu!czURiG2D)=s4E|5GpFȨ%-R6kC5`PD+DN5ѣ"`;T>j `#q((: 5&^ܣ[8 WAr)qO_y9LWCB\DKdJbtdM")v;Ef n5$EH>xs]1v;K11ZiwI?VJhv !G.+23e p/Fa(AϑD9H^[Ҙ& FkU[!J櫑%Do76|3̾ 5p6ΆNxKtjGJ 1RO[v]ָwZRkQ5nDtߕ F.]wTKD0Vsև!'`,m[FKtw=23wWfݟӱq-RM."Am͛W0/?|g$+:ȻLfx 0>2L-<! %&5yO'q;B)U N jg0<[ei#JW)(PvfkkQq+]&D%Io͑ DɊ-7w0k4 cK :m?ӔR0IYђW,vDBq\Θ?1Ot`+n  QhB^f ږqV$ S:<1l^J*JF5\`ӃrF4Xp-C0 3vfT#ӥ/i;/JaȨ-VKZCnTb2f')ٸDe L&o=8s DϽ?.E>^Bz\iws/?=Gogs{#5~3,QM7ُ3(FDuP.a{ZިV4##lFv5O ʃy PݩaQғa{]ȕ|b՘iB`ӗ<)]&w G/ZC ~ nΕV_`B{w׫?{2-:_"nlJmta&r⦶`j۝ݯN7]NvQpJ-0/L\"pQ4"0Vq+A[R77j$BRAdAQeTLx~ieLkOL;ZB<諙mIu}J?3ra||0osջC(n>nϑcƩ~[P:bZRSY?B R͔ף_fW-r]د; h zKUfM/)ΙM~},m3K- j#su$,JqȋuJ=ûٚW@pX֘rz~Bd]a7 &'Т'pg{g6dɬE#}Oʹѽ׭Zhb_ɿ[L\)񄧆FOOBx̴@učםK&۔2ΈFі-K,\wK_5.\PurטQQWT%0b0\I%cxemjF$>b6FAzc~sCr(b}FDT¬DZk,fny+9l_N&oq;**?ͥڎ_b$&Bij:e$f]E^m]7 #xFX|SQ7naWe n^Ī,ˎCa[L$ld>\jOt1#Jc~7/)UW)؝NP{87U_KOS.%ER ӡfH0KjAwkaG60*TD'WpkU[ ϳ5{{Ybі1=}E#9 C(2FpnMh-:U\ ٱ?.%!dE^q&7$-6hx>@Y X.FPdB8zA9 "h!HP X!k)rkB`U6$"@ $=ARC1 7*(XBl@A+ec+][@j$TYVuY{L1_ 9B)\;F&90j52 ,0бHJ$ PTKzɊԠGr& GBU ֒n "' ("5 $Ѯ o*7ǡZQ䕃RJlDXDAđ$`\KYP8Јv1 LR,FY=|,!pmP ǐ54pIRE"Q#!T9m;b2R3pcVB6H:MV h39)K(u[՘%x‡Pq31gd c؋ J5pVnr6} `KVYQ"O)J54b-HϋTyi[zogUgłO(]0/n Ÿ>qtqڷw/j >#?~ lXOhrƘBەo '>Mjn]he@,J F-vok~9LfbE]z71$,5E1`qK:Z F{K@8: #=px~FtJ\_~ZP&Zc4 ԨS'.M&Li#ѼJyD*UY4ˑN`%T ,sK ׌d0yH!A3|Ȧ@"-JpY.ٞ&O6=҇1/sN!BHA Q ց}& 0ٞTzv= 1ܪe"LQ$j" >-9  NJC߳"bYʍi5*b%Iu]ak5vFQ^$[e#H Y t.w8c<*k^R  $Hu1rrG'At2 S%OU1XN"+S &6-•%2=9!4/3\Zj{fk 2CQL-pLnqwYS};3T*Ho!(0,&6*۱Th.JfEm=9cRDPW.hr2;8F`@b0hTԁ J  D]fN>APw$r:j*})THC/:e^EDD5FE" lTA' 1uҡ'1*4sl-#JyQ XmmT]F.8tK8 ):h!$׵s蔛TaHn+#AE%hFRؐ \)m pCƒvyȌFmݎZ}(|@g\_|@IS ' ;U2 @[R=6u$J96QA`YN-^'"%dx\|#Bv`kbZ!X:̅5Iǿ0zSpZ lv `#l(I8%`₍BPNٸJ2´2+(ga(m(xtR+yT{ 8J. *eU.6= o1,n(o#:E)9|8|U,łM;FEjHϙ[o|Yv:W &]¸ڕe`'|܎8//?_,w./1syxqŻA`W$pBy MM×@t聣~*ˋ +1*XSh ]&Hhr9C-Ϋ~cĊ3`/H(ׯQu߷c++4ɒ_LPFkmIX=6詊%DUr)PSQ{:y k2o} MTfY79Uxf(b0Dk|n]T@7{E,gRQP Ie]X=xdz]PFn$̲d wZ7I3 \*+W%"?o^DÚܠT(? :-9أ =0 EE˚ g3o;,yFhp;9.?g7o~ $9<oۗ[{R8k=u1Yk8g5Շ9~}>ŻGL;ɐ Gs=ϔDp9̈́&D$5g8Xh1ͿgR] a91ه鯓 }!P|w?EdMk+lwKd]ϓkcף/4XFv|WvaViWZDk1^*gq+?j5 k O()ʲ P[mx6Zp*B6x!B Ѫw3D3yD%!x DZ≘3K6"D,Hi-0GB +Jx<2:*a\jrm>0ɨQKm@:RƒrG.Fք: lMCC]+ɷF~} r7W eL7o,@x *$4S>uo)D~Y$a'0~AY<\YL%a,B ar#gQPEi eQz [#,4VA49PM\" QE[Y4p09j").RnYi:;uC7NI_i+ss8{VH $c( 2FdV@,V(?5b2?}sV+ `JPG9c"F/u `PEPP#3M[#.^6saA=.5N (ϵ1[KQDKβFa }, yRs.[rF 99sD.n4QqĨ y:[܄*i Qya#b[FP9(cɌ@#*THyHw9So_ny  gνQZWbaY=(Gf`qnġ~?ݸ`2kEEnBaG:JԄ}Gn\tD+!hI!9:BtWK]xz}[WW'2hQn,jOZ\!yAq!/n3)d&`os%"|CbҢZq\P3(!KYi$Bp kpq}+zm{ %pUTfĖ%L)>7JK,Lt؟ ΪZ9~%t?V=LT1$WG'1EŽӯE|\wִuϦfRKOΖXaͧ~(zr'PRSėҁGKG:É:9sIf8GdFyW9(M+J&1@Ee8T:䄚k?r*XUjqJ)m ;i&&u50ȳ6>1H׉'p2tu%; &]]b)_>Ɂ<&mg^~Q;.YOg_|RT gF>]WA!&3lDN}y9?'&Pc+S^p/b䚸G w+;hRJ:6Lhl[yW"idkm/G؍ݢ@-6!9Gdw=3lI<+CSI!;p9M s(:(Zgu4ϱΓ<"jߕgؗ!dyGr?hȋm<7jM;lrXo kB( 1ǷV^7oYfUM3)vMhwbh 'YS5nu\/nr7g1jNߛf4_p ,n B4~u5l7nK7nzxݸ]WmO'{FNrm=:նx\u/SwE4wYBKx/u#w"ApL_hlײQ(#l<3w Zw#W?(~=:_D7}ۡt~unOQܪ1MI~3{ɛnW ]KZtفa.;.Iᨏg %4sݜ1t8RKޮ!ü8@WNi O5p> ?i*`{^;3]wjZW#wE! y6e "Sj>ː$o \[e^Ѝ\ffȨ,Ԧ4 A'; Xdx#5&˹W5{$5d< >M{TgiφH,t_cf<]t{}V<鲝,<: cwC0M0DH.~|sw(M6=A_kEin͋{ Fq0nѻ^W9'nHv @U/w[$؉NWbF0(A~^QYWڬbMM1m.G[KiZ:=ujܛO@jNI>|Å/ff>y0?f=I tFfeGV?.~zߢQg٠0ǏnLn&W2_e$*x0 4rqbpH{B$OS 2K(?C(O9See-ʿΦ,,7w||vogY&(g`zX}*sVI6-nX(baϹɎ9>g(xRޮ@ЅJDSܾ| VGPPr-yHr6~c:"*KbZ4!o~ZK o5APO\0g/{}[ZЬ '-/%FaOUj]2ia9KT8;b(%(ķxnҶ5RkT&6eq'L+A)ʚ~'6210A ni`wI ۘ6*@ |J॓DJL la 1Y} pȃm E.8~e~7$08~&m~;Lm0 ׮Wܕ˰oѳ-}\w/Wٗb|'o<`V8qV8 OS- 7*`%+NNFƧ)ň-uBy wn=]wiYڻotmkvڻZu烝ưl:{.y>UL'Zb; I©MI gfjr,pAk ݶTn0"]թŸ>1YSyv1gdb·n N0IS)6@JFE%'1 0'kSiO.C\ӓE9$ŢS[X`zvdbAłS)*tIHnykyX+0J%  84:Bټ_К`0Dׅ%o,3νO@9e{ # ,ةB"IpLWAQ0@I6)uy+!-!˄ |b )Qn)CZ3X±)-j67V3myG'lł IV~֗/f-߫_,SaF0KL3ߍim98;<8MG5wy;w ~9MF d#yak6fG 홶oHַdbPLnQʳR2G:>křQ7{T/y󱹝 <ߣNS})P !Pò⬃Ϛ`+: ٳZfW5g;A]}BZUC\}0iZ`5ܰŸܞ@J7K LQ.hq7[MD"pcmŭ]w˼CG磟%#>ݍ曎O>` >nwbhN_y/\zhZtS'jNlۥWR.W8ToRq A ֝Tִ[xvCNECxJ~zBIDX[(eX'tc[wRMTUդs`Jj:8Y4$K jP ʰNhY-zYn%[r,SvSH>hP ʰNh1Zeny%[r,S >RwB1(:c:j%mMgnupȉhO vӲ-2:1ڭnOעnupȉhOU;i1gvӔ-R1. hY{!fe3Bjg -|I;1(1q](TlzҲ[ r,S9/._PbPuB8FuWl/-bvCNE+F O5iQb^ g,#"[z1hz&${ H/ci2YNng2^.R} }1_ P$OOc߇L<7058Oaw 5E4%TcγoeޅZ0a MuLb"fP ŜBlVD`pxN$ (MeP`5 eJhoa{@;FxJ]aMG)O!,BQIRb|"(S)։pQũ3|4Ή1<3P  -AQw `Æ"K%^io2Up'YC F#  p!(@/q.")G9'i5# VE@bM5a O!}n\|q{?7<8rEoy̜REfM Uhxڬgfqv3M6CKgep{ˢŸ1~#Iv9|1)l){MT` ΀tȥ xV#˹I$R%8 m$:h҄X + OSuŠXb!<AqԪ!gyв ofoJVٖpb͕1o@)Yb4Qe\b{fڷG>LF|ty ] W t4;f2“`?\^ik6cI6͟*.%`Lyτ0]U>l(alg#ώ\K$vlљC<ʁ?ssIGHp඀3J#!KoyJHMD F}`Ԋ/WrÜ8Ra):) e0et} _u?f;I?̶f@_ܜ;YL1cZ]8 h2 w}ތ__;g|?"2۴ UW Cua9B ɌE9,woX\ƠM eH;gXȈNccL;nS=7:mOf߉4]NCZ2qg0 1N+oJ/+v=@?54վ6.5I1?F*[Ғu3=K~dSw=X~bvv/o&mNºY+MȘZ;;xay?O¿9a)jW4ylNJ!.TJDTצL~b? X|tA 2 9٨x&us}w㟎A n 0ף7DIoU|"O2𷖆;vXfQKѺi:G;@}JD+%BKf?Tq8M }JI" 1krH `VcDZ \(&bg`.o(uILG d0j0sկ˒ 0߹Ϧ`2|\]XY ;gƠ=`Ikk>MspyLrHIy9 ;4+LaU.G {2_pq=Z0G Y*8e?͠fV6IFQ pswZE m`f  z|̏c!דYYci`﮾\&7 )lnIWo`G:I+׼[w_=?9]ο?o#9'KؒDmՆ?;gq0bVDZnb(uH(]NܫZ|a7vwv E276!nN!*5Ěp[In"~$(b5˫_fA\t1@(zRQDgڡ50yPQARg$S$zL@Z}E> "6,@*VGM0aQӎ #JkCсCKXJE ĽFU(TAϢ-5Sm0\N֏ ׇ?26 X@H.F&<I QC*I|VPJX, b;oF-j-'4SvD_] A kNgQT1WAN\;ۙ9UjIkV,lCC^锒;Nn Z{gR  8O-p[АW%:% Ƈwav+ Sf;n=z "bA!\E7)I$ֲͼS9v'}IC+A2f%\FQ>2ʑvx`Âr\{X͔S"Dh-C]ߜ\Ks:OcݭY0?'}g)W{r.F}G_d2:$CJcS~MZH)I)b - }r{B_DT}҆_Ya垔л3uQFQ4?=jܨt)T!Z-6TO>]ߤi>t\ٔQhnΧJ.V/a3#}RotXJP@L ksDQ9ʷyvSI<'Zh`PNWK\1nEe^(ϐTzZg-c4ѝ4$;rӤ2XU;:nџϧbw_^RA3iJ>yt[HtBfudVbưOD#&1_  Xajϙ `$d0HzS0Ih]Tpe15@ףlŘ*[fZwP`o0L}Tg)nY8hSqB P5GQFD@D{&/RgzpHǫ"qRզʞ1j8٬?^UVC!9/^ !㖄W#'K)]- $UFZas,NQ+T箎诼~kS,Dc܀!rW E[s^l.`oVbw(1x;r)vMg,yǞe%b=Y KXHe~*.djIn}9ba,cDZ Ajq9 *^*1EiŽKJ1[j S,z I.YͼXim8U RK{y;nBkKwh]{Sh0,W{rȥQ.9æ&hٹWkYSq8ϻ,rYst EAz= r y_W_\ )!%Cd"$KE)rYޕY D)A|faIGiB*.J)4/>^b[Ҳ~yH29rΩf"Q\8KMƋ#L ^K<1J**m(Rb-UIJXl( *WZxE )CB-t41:4YP6#fh;ַe.|UX&lH `-WXP;,LdLF=2]S*X8 '5@֓ H1ʖ!7/e{Bym\D<L ZIqG?X,Ř*[f[8*e"L;A2#pt81%1s?cJ'O([$V,Řol+QkI^ \Gؽ;ePH {R 0>H<_yC5?j<" DD E ?w=T0,T+Z1X&G JYogwl"lAa=j_~TL'r~{WWXD/TuY:+ weL^p(`mΞ,.Ϣ rB%rf~*ͯjg:oTNW$ۼTEPwy!g{MvP׹ ,?ֈSE򬇻5IaݭOٷF̟hֈ&ʸ1˾;ڼ =1N_9F<{"fw 4/'iݪ~qMԆdBsE͝رwwIyJ(A\RpOK}\R]%quV;duh'D@ ÌTdUV9LʾEE|cS4(b.@UDfApa AB! <%)ckHNk'9 f܊SV:@,+h@#s,#v!:O %en)tY(甭"eS*  ."E`+_OB!V*H;̣{Ө DVrtwDjy͋7"DЎlbuvC]\+]hgHyxH5! 3;6ϔ =@nt~m5`!:r\>=D[dgxvw Eי!C jwi4Gjb銄#;,`ں`V^'W"LP%NJa#1[a׎G=nZ]Ÿ׀_S_asCk O:ĿR 靂5w7_b ǽ|] [|$w15S;_Krk~3(CWjkI R蚯Z,E_Br~UB<޵6r+beql˼on=8@O +#KINoQ%lEILlX7|qiW`Dш>FG)X+ɪظ+tv/!!=CBY1,C&C#&^b6[5:\ؾx.}v6?rҋZ* 6ӹ+7zcX8x`6~a6Io02ZSxlc_BS<)C.&|Z4aKEeW6<"TB֧C1y145 F EkWۺ2^r!mZLUPeQ[Z\G0wR C2f 323C1 Xm ^ 2`@‰٥ێ!8hqdȲ>{枀$9!NK32$Ɍ(PMdQT ٔ@ar`nm&_ Y2(jƓ '8kg M$B}fI4`'>_u]-Qr$Sp8 ɲZp d/) l{F*8KȈ!6L!. ̡p:J7۩n PC|RYPbAxEo]+Qѧ4MN0b(!<>a.{ da(^ՑC Ruk?R'I[mJ52SLNo{+&2x@x !>dRHlP:,/2zX)3i-u#H~2JGfdL2[(v~g#(BD)Sr#!VzoI _w5tnǍ{_ވ-iG$ ^ǘ+,0b'GE9? `"~dvM'j]OWv tb <ָKc|>ߗbk(%͉( S*+*C!fmp% E4MBȓD'+60aiu sbYBԗb-R0Z02 $ܡ=AZIh#J%DֵdpIu!d_)Sd 76ggDS!&ACHs ϕдȩ3* éTSiB/3S)RnIHё&}XS$Ԫ6(tozICQ)fV+E376`=*5'FU'ndC<{812OWaQw6@6M)6<`X ?u3g9}>ͷcv >-m=-i5ZGXrZQF2!w7~l'7|~sӐӏO̐͊M6E4~^"o7*v' *E%h^* D,e4"$͸WZ+GCX>M-2}@sHB=w (ZP5{Ήڢ\bg o&["8q=hN27u#@7_c&Z;.9)\KGs:(cX`,>逾L8&֔R.~?i>63q0ˌofǠ[kw :~#(Ͽ}D6glWOc+8˻:nю--p?M&׫U<7/߾y|wq(CG1nMFՀG:p-"?+%wrbҏ$vdN1gG4N׎#WXh=ENP (4q9׆ph pk}({̰FFG~mn^n>4zݙNC`h^P4c^~ĬsL:bfykfxI0DZdA]!K]DeQ׿/8FvL ŇTo>:L8 s|bP]=~iƺd1{n7?k@ܙp0m/N39q&iA0fK$2Bb Vik%qksʽ~֓tYV< zwL>¾ڤ7i,_6Ok/ns&" #R9E㲽fPplnaKTj]R|t_G 4jq[qe'_+iK/8joq#ЫFץ3NN /Yp>}?Ԧra֤w_C`:; M+م?4~:GoJ7'BsY\@_Ŀs{,It Q.JWsr^?>5)͡vӾ'> %gWF{Ź$ 0LBΕRy0ّ9vu4*)~qϴ8atm@ ĤDv|6d@x:g: 1S8s[XW08[o,\٨dYn囵tgv[H{.];>¤fDPSa :؁jQ G/p<+fa@(c&TK:odE3-%g@Ҩ4Wíh+ԐQL4PA@'UFLX5PSC ʧGTk.ZXxb @Z}B+5Oa: u' 05'މbH ~A #m*b ߃CZ$҃s]1?ul0 c5A{JSe39rŌK*-R{j*NUc*aߑF .n:.]>i\=JceoG*0zP.O${ZƋMHWZJNMgFk{sY<ɑ-<6!>^xv6g+Ÿ=̖Fڮ0]|/g^GtBMH6s-P]y<,!Il7@6d?[J[bϨ9<;>Ak ip&̕S!L}6$; IEs]/ɮ>N/l!fYnkOhoEf`U~n>^?x g*)[X:zi5-/X4U@j9߮lVx5N[S_gn~z@LZ6偯]>X/0nWׁؔ6gD ]o6-ޮl7QB5H#h&EC Kem bƇ^*.|@ Vp(#܆}-Ȏ0< R{Vq2X!cj?LX UT+c:cC]۱a9LA0 QB Y+ehh(AKvp[<cۄk"Lא gNjsJpRKugZv.j"X1LRAєg'? U4Rk?G׈]:m3y:[ Vp{yx1U[dIE$ޭsZyd5z[ >ڨYeQc'?M*Y_6dE fD1J$S<0cOs)H\?Hz95 b$?p 7:;93cK Qk#4ZkXgT^'?Q.yT#HJDghWUJ^"O(<ajhEE]%b|WG@vxCUfggI8&Jb%!b%!*[ZL3L F R y.5BV1El8w~#{KDx(f' ң@ >/tJ!9B4UZ2b#Z[/2pmBADcF*! (?%6\5ƅM4< mG4r̴J'QGݫiHJ%G֘GLaiXlgp^hfXg522&6/19$v ;c)+dbb"e)BU,jVB%K6GqAn\W/OX太!C`TJBi^9Y7RZhB"52D9 B#=YaN*GHJR>Ԧ('S U`Urf 2 ?'B2Ly^p }fkȀ)ɘǠI3͖ZpudG,&R 8E(OM0B2oly0S{U@S / 1o#4Qqʐ$Ÿed.)PRbW>34nsF0nh/n#j\`8IXhԂ!X? NScCb\l}];ͽUSyXMesS@ەfp6̬2 f nύt)ҿ_94Ffj]TB@7 ,XA,XU"Oi4@.02QS)e)k'M)\C))zcnJ!NN)R^hnc[R5>?Ѩ'D9bO r13'mb4h1JċI QdB-GznZ tu1^7ҏidnَ  gA,+o*{e]ШY̸ wa[i`5x[=Z?+F} y ƿxsSjh{et<~۶փW[o}>_Mn_RNvۿ捫eY.,g޽mfr|4&~|ݓ_c{oeMؔY5VF:z}n^WO4{"mYe)nײ܅#NSږLVn:|(k'ip7/{_vŇ?M)ɭ*WM'1ޯXrlNC_:fu r >qw?\ēvaok Cu=j`:18ʃ7% +> ߯a83}~"fBkռ[6~I})!lA nRp*ZZG]4S F|sG7 ^Ma(BJѷvwt 0߮|JA!q+ H'_8G?2bqkYY+y'#( {LLcc2(}qsa ~6y+1Y[0*h/}Ud!9UZf7AE.|v̗$*%"? q\*J1 4XA$D*,DH^[1X8~DbEPYVNʐ.@ װ}iDaQ$4g(-zC1  )4R>FjKP!rF1Ud\?&eYcTiY a6i!*sYgl a %*fMX0AM4'UkBڶ*fHF0'S*TZQJFۇڄ?c1!xqNyUs柱Cţ.pcmA0ЃqbsI|PY2uR3fNT':HxXPsh$ (h ̜(,*U ! MD E@3B$nNh 2K d'eCm*B#QFUU 4L]B7f23b,2-[A`()AgM vkgQzAD=+֔lQ9=4THh*2Ovl!W+<)E!Y75pa>$#IFBEYb}L4n~xVHP"}=z2@&^5d[Zl]N䪡yQxVU⨄Z{1*qԫ\FAMYإB 5§^ ;e0ӝNԪMok|=uLr~`aҔ `5ouBj7/j DN\jĞp뮍O~,ی>^'xِqGmYտW{i 7~±rlD/+͗c~22.uڸOE%qb6H)WTcdDZt$>B/MM9|4ZIYNT<4ɦ)A,<_..q-"hxlBH`qZP( >̗)"e,(Zx9э@|4FK7s`)26$WqjS ,&K.vyT!q8!B"ZP|Њ|ZN*03hh3h0Z%4Vf1kg5*&Yqd(ڣ7OAP1^2T ~N}_.l"|/Xd?7MB?pGt5㋑Q %ZnΚ5Uxy#NW,?Rы:cbŽYBlܻ.1t8F,t|[m0+*nsbTĬ QlQi{%dCsKr N@Ker`f]_oU)hDmP Y{l~)%6ʭ$R+B.EmY1*wN"1Q)ؠCz/k*^(v_/|Eы AD/":hvCE z D7N$V0C}pi$[ŐRPV8w!ГȪr8?T_Z,=ܻPX7 0Riui$;D gj#Ix*+Ʃ-H-@i!R'ZYt)0ǐ\X'?FJ JRtyzyIXHSnQ&A|&,B l84%6s|I%siɈ^|3-/](ses,c15JD̸82ڊ#kcU|"q*׽S,( IC#3:.IN0[n-STzKj^XU-8peN6x8 rMi@VvVdB8Y'3亊#Wz3xtOCizuq!B=^ޕϤ+ih<HɪS~Cc>DmT_ dZrOu #-֫&5@/&ĵиZʯxViEMj+؊rJ7Hv6a>L]M *r+YnNKM,o.DX2PR*׳t&n4˜IĶ'"km#G/Ş4,&`w/d2f`M2Ưcv_%˭G[lZj;N$QU^,+uS%a` MZzn YMR/#e;PKUH,#^6*^M$ c@w_vKWSt"~Yt#%Nꁟ=o(rYR& 7-G߽}r|7W?SD5s՞L~lGt!AfxyV\.ߎnG/-~v{'ц#"ZjGo\Q1Ԃ)'g.DqJ+)dg-G GZM`,0b45_e #Q|a⻅i=CM%*GC|P#g !"r|hRkbi" eFoAť"v35a\JC[ G]St)}cduD&5v[ Ij _XRbt}y<4IJ+y؞M.fo7Ʉo <›2׶t!8Y ~ic\ކT#;vPHPefB \HNR)49AVEV#ukxNZ)8JFpS 9F>jQ(r(HNV )

|LFooPjy ɿC3T | gSGa 9)MJ1m[fw8&KڣGtL?MkM:k /"\in̹By\sz^N0VS伧rK*B߳ Aݗ_oŨ1zh:]7֐9!ViW֘7z͛W?L@U_PU:)22ʤ>Q_|h"?64O''/Ho4)4A:+Y=w~SzWs9򘕻'E! !{ g9U.P̰EX|v̊(Z݋tݐk R,4j"ք`ڂ:WxΙWE pɂPj" 1PcdBbS8cRJ3ۢ.F,C6T!]ѹݎW&n&r&1u Z_W90D)fC!H;LTΕ' )g:S]Ջxn/",̣nnr$aT;^ wN5EbgCH9)Ӕ7FaսkV! MIh_N*YWx"4]7¡GGAzJSO5zJJ4?]C9M7].RTdPwS(b8!?YCh^%ѼJyD*ʭծZ[l(V%Ք:A)#ӶH#*jpi4K\_ ޥ vA] [g9k+ bKX8_&"SPfvTqYK5ZOX,<@y08#[.1PQnjԑZDЎRŵz-UݮTqk4CI8{nvi97`0!(dtqN+|Q~ê4tepf0J3~p0lz$Jmys}rrNp:to 1cUe3ӗ= C4KX6;rS(Uj$ϓ9\B< RNhݣ$WeM} HK#+J'ujsUB]M.B0Zݻ@[e@ u]h+RwǪ!v-!9I, DiUxABZ CDQWqD]SnF vgs+f"gRF/#1E4JiU=/0^!*nL'W:~Hu٤s6d2+W876HmPADZcu Zj;Hypl#F*;ZQ'*FT8.h KtSh& k r[8*0daYP$F\Uj LP@2r=7` w $: %9cãH)RFv`JJFa )m,hsQ1縡Ƣ jXMd+ ~)(c2r24 \Fz0+uHQ160*)+z968bZ+ q%"mXjy6,{!j24j:N G.^*9Py^ ڬ? 9h!LK(Ltb( -y~4ɫ%4HE$ .IsѣCnRnTrȒGÚ2OdKgZ&|$&LYXS[bh;%6k oHK./^O~ @k4h! $Q,i" TLf7q@;Ϧït&}1x0`QvKC:8H#)5g-Afqs684ǘ\Ua_xb&/h_3?sV-xُo޴Σ%]C=aT7Ч=C8hfK1%# F`|Kh?ǪP̣5m@Zvd%|wkjǰTSX"WDTSܐsMQkQȄS{Z!CInx"D^9kuCBA.s>D!2CX-,6P,s4ls3Ra' uCch D Dnoz-sMF753h{P]'شD}z3 K&+vH% g0wmX2ؙŀ,6AIeҲt<%Kҥ$V*+NMs!y8M%=[cBЅ/q5k_jby8^d7,as8zS-af˅e)exY<.)F7eT"LΕf:[Y$]h9OXzbC3 BzͿ,AVݨ7pTKP X.U C iN2~|bk= Ce!c[Ij;޶ejK݈`]qbTxq&Oĝ8Z|/ WTʒec~ Yqj2H,̅7ϟ)`9E#V {J7AGA C9\ƃN3?a> !('0_Ŀ5; ]\n~`He OU#>EfҸs$Զ&ƻ;֍x\ W;<ިרh%^ݾAbnXh3鹧?lXA>Rߚw ]裯W]T @jnX/*0}' +o]gi&ZQXl8y͋DoԖZaF^no(S SaᎩ.4& +#+ִ̆Ͻ$L{fd:RoO+Nɝx|7TJEɷUoӅr/v0UчK1XVӏ%ٷE< v[ܮmIeΆƵ7z$yYƅ1G!;EvjX*d/Nh#j:La,vB~o>j;U6J0j&@<=ìJaa$)#YgR0hXr*P*3̨`LCf\R!JSC!'qMJp 7IVz t1{d31{ GOkHOjXSFHH%1h#sf).LQ2@2!'3ĄaFX$Ls X\@Rl >%Եw-7ƓL0eQ5߻]e=:B֥7 8/!v??,u.{RR@#yh7ǯQW}9\_•b.ぴ nI5+r$l.^#B2?٩jY 5}1 Ŧ pEN^-M~*9ǯypRb*ܙVM5)ivG(aSbryV6H "d˩$5h\j 3č?oєb"T|ֽVSJܺZqi8emPx1ᔒj]5rtLHv-˺mѲH5F /6HMN[d< 鼖 m:άH\8tI]t }ʽZJ 3[baw("%t7V=.iyi9gCa܇ru:m1XE6J՞f^oK̕XӷQ`>?A}8 5DDPMϣIzqh5d|O씝T}:d1ѥFG`5]‚)`=zt6vl2RfܳsjARQ ,$΢0QTr(7#߸I⒂)}МRABQPg%}\i4gK>+aUԋ 5΋Ygj]iifw\V{8].%~EMGu`WF wOU2+|h \ F33P%h2=^^7)z AApivQ<ӧoy}O`(nBqt}ﯜ^/8:{.GT~/v;;bHJto'פuFh.f0ĉO\s.bīf]uya~yf9{蚂 ,V))°-P9]4=^g옏k~7Bj׬@e&$b^4q!${k3Иk"L׵|=>$^᭹axe$IYxC7ILr-νf"PCcޜQHZ?' }ΥJ% x.|fydpӹԺګ"FL(S) (idK)k2ĄD c QIz$6H=?nqy# ~^llq !P #!9cL0"QRňE1Ki2(GPA/G e ZҼķ28Y[`4u> br6}ҵ'.eYbF KU y[Uˁﺃ0PW!ab,k{* +z)=OQz6_HFnW3n?Ǫl/`spz7q ʉ_EWnx^¸SA3Hn"*)t'lE,z)mŋoEhZ ν\BH(܉WJm8yݷeVTUۮ ֐4SmVuvz) ݵZ_8DiCHT8.]rzu 5Vt7vFI{RI:meK%nc]mCZ&)89&C}"7VJ֍z\StnPq(cʪuHe4 ,HfAށ=wokuٗ\ qjg+4[CޟWp$t3o٥Դ$hBty--NνK+F 4Ȉ3n+*'" cKG> i+eU $GsH f^k/_<>lL"~3)YpHkW&]_dQe\73|q[Yē(ԩC HPK cY8:Jf[d֞GцV`DW*2š~YUsK0dzUyŜ#Z"*xӦ~ʾ\cl6)sXqDV; wyNlMæ;F|qQ (`ONYG?ڴX>VxcQ%򛞪r¬nv)F>N5Y۾`"ֹOa艗k!!5%l GӇT}k!c[#)'`@׭U[~V纏Z7NnV quԞ#Ly vhM(ǝxYny6|H;51^^SUS/q ҋ\HrM';aF]?34r-ne \ZZ| k|\wݑ1.oV^ !P #!9cL0"QRňE1Ki2(U@$Pb5 ߔKVpPxW7gVyz$ c0#)*ӱq=Z֥"$# SzKb9AQM7"FL(S) (idKI#%ޠ !0Ei֗-A9%?JL=Tf>!).U`:TAnS=|g/J4SsnQn=i֓[OEx6B$H3,$K R3$!RL0 BjQsA*IC"eSs-6x͔bdqS+6zu\DyoJ /iӯK0L;=9b=-]<'LG4y%6.ARR]50^u.k_^G~dy]N% 8¾Z_@D.G$1ȸ2T& AqDCqCEaFcL,K ) ~Rv8Xv [;R 94Ac׫ZV ByP⃈F3(<,? SR)x /6ݝJ΍Cba2Lң"̱ħg "`.(Hz 0 `#Rz})#8mQyI$r/331J|fi=ijd`}=5!! Cw͞0.*׉cmDQ6pJ$t?7Nwll4qx ] FDY-J6JQ2fN;Pl'n46I&X7xLvQ&m*+GI8ZK1DQ˂$ji*t%!k^Qt+$ b PD m9R1D$!~0@19+Oy q={Dz >.@t 1zp^q5Da$}B {_ PL#D䛄?"fi7&IFgKYAջpK#3[{VVٛ[)-O$S3UT#SPA:>> *՘97Dm3{LI{:w+gZ;N4YT h*8mi+-bmҥ3_~CKD`8Tp/}R&2xUDAt_֛VXj9 8{&t0t.'E ;djiBs]-'ѻY.S_vsL/OJFB9f˕gx Ě#ȖE4MbJPrDSU#1{roZr7FR~;~W8ml96X7ldYɺɨ+kJqUƛ`a #j;VWbm2Ws%Fhi {#vVH(JSvݛع$@ nCKoi#-#_aNSӯ"/ ρ{񳊿-πZ~kK-萜5NG8Gc:\سl>J. S8E\:C!aG k /XQS4g k>f>$f{ZYݚ[ˈ#$Hg|i9ijED\!Nkү g؀3墏P6@zG l>bq{)+)@ FJ%#+R,FYRIHGC!A_`f*zF>=<]+aZBO'!бy.-{-$Ќ46} VWDjMB+MQZ/W6jT|m;y! JC4oS9.nR1$՚lB:Iaƣ-Qy$E+t;T0u.Ct̯]@[7d-<"TʻQ7˄OM.O9ԅGisg:OuCKwg%i_=&o}$jZˡ&~=& Eu RKIzS84s#{%$([SPBoކL1RZ[9R3$X{:vm`E濈׫A 6Wc6r̗f־_͈>*iafzGe 1x S`\UߑJn:<eJ1BiN _ˮЙF<ޯ7o潦٠G[+=ODULs* Xf.54cߙp۳j5yYb'rhdri[jb?F{LI{:w+8n&-.5ۢ 6ffcXMt//?9zOϹo6GmW!9(z}9MQIûnE*2EcN7&1@#4r7{9;vҋ2`x$Gzx$qFܔٍ9U)@~ <•1Kl9 3;e)d5#ޯw!] ;yXStoڰrBZȀpbۧgUa4kDN>|vq̡{_<:m9ĽmÕcƲ}12Gנ߇*KрoC H1Ğ0skR#kTN(7^{NT&FA(p"(<$Wd)1ǐ4I_Bѓ倓VS*bda$}1$Q( (IZ "O/S}g8n 3_jIQja^OQ *@@7;6kq51f"R!#KJ "$ L+1#  @[F 6N Y$/&k,x)MUc'Li=ZJ.AOߓD 9A2Cw͞p'=,U򫯓7Ez[@l(CU^kQ5c[c/L7fD[FИ1Bw/a$`ɏۨi7!҅E/T_?j91E~]Myzm>򟙤M3\-@l}ţ$uǪ^ y6yMR xu8^|(wߕ7sě`18ḻ3EŲ^6y4}!dÛftoI28-!Y%jvo((?">FPyJ LCyF/jiC̰ j \D—WR C=+| 63TvcHR=[GӉn߃mraFH7sFGOP@3Tg䬖U嵦nxҡk;u%?BwU z'g)/J\Q=1q,SʝZx<<PEe{r* e˘2D00k0 U D|N1_HIOB%Ml%h,/¶b?{W6ݙ> àXL7l/UiGG{<IDIIR6n))F~_fDt U#:s u6"`-Jv]xS]\mT_!FX+oIxej^%{ޒhC!}[= n0lv{'lݙd -+0⾙XU>Mz9J0雳:Y[7NcytPQ̤E9rܷ(N6fmyW) ossz0UFW0j]FN7pxԿA+R늋+dWg[GO{WsVXcm.KkR.2zpI.tvtg öɎHg^ ]"Xn/O5F$=gGhcX}V[h1*Mq"*~) 5G:$1F("(LTB#XłQhv``Lw{e\MПzǓFn6pFl'fCqe|?^K{q-q`ts{0s XZNJUZ9(  # \+ zr&y 9GI@Q$8a @#'@K0HP&V`Hm6 t[hY b$b'1(Q:$a5Ea `0n&G;JF|l|9$JikX Uf:Ǎ{Ká!Ictxqduzz`g ]1(1f&cC jl/ӎïŰC]xJ$3bU0tX:МGN$F0M0C'=\1&rNR9S[u.$I@jA;Ƈaxߔ)Fb.V0-\ltU&za~ŁzdP͢l;T< /ҵ_Q+YX֡۵SmdX$Eud5N,EJ]\ YP@3XpE \L, %k'k'99|}&XJTɧy;\QKKYR%&kJDЩ$γydֱȋh8b1CfFp{D宴sDڦufQKDZ9F16ZyC+_ߞ{{NA%  xzPi'XQ:feM|Nۛ7aY?밡Ga\<8L kSzۮjzʼnդ9 guhh+8פj= gh8Z^uY`;jv%A]pP K|KjwHaZɫ vy8bkʽ9l]@r] `CT,U Ӽ GFԶ;KL7[T8EnP4HRI*W̜bx;'.cnGok;= hB &UbcᓃޒA92W\Q Nth8S*p̦`[)-̪V-%ۿGZfت "+MnEη5W6+#4P 2 Im2^ *1amv4D=++FĞ/~LP2vKjJYS+w 6c^ ׸$&Hۿj>z̜V9K3 D畾],z]NhX?QJT5h+֌&`0BA(u('FhzMW9:r>yBjjF^K0is)ci9xENHd'_4LD{OL"*ީ8>.t~Dəw\#5Z7m?.dZŠuE(#j'Zg&B j5;9 %p:K%4ڹHDL}K_2B )#_ʤX@ '%u|.Ool9m~w=~[UL_.Fָ;z[UR±WJnKh7m\QNnGPԫ^C_lg0d-;P[!,ohST9ŨsE4y砾ۖ3zaџ..̓/3o; N+v/>FĽ4.Ok| f*n XjjeC"K#]dWO}@P:B_D!1x+HpDGXjxt绢:2$qG Q:2UygkAD;q Q2Jh q@(&8'\8*0frZ~TI~<q;6i LSJ)*ZwwƦgӎ#Te]_q`nO[Xt^Sg=w߂|xgF '>ygw~Ip[Iuw J=k ܸſvte aDrہ(aQ}ic[ Q#; \KT]ٯ<[I z] yqQ8K$X_EAbٵw[}n;5-IJ{}^:]cbYQp㰋SD" .vhN&w`OetT!wa_uY{Ұq+{"A+VytP;qٶEml\i`56[4|B™wZʻQvV1R$5XɋqBU'[D6HUNLgDM樹BW[7jTmZs12newBƉFDk;V{YzN<78KD{^Yr[7jytPQٕLSДiqD#`Z4oV,Qg޶[ NՃS^!5P6s* F͝yT1^bZkҿ^_ nAzGibIhcb9EWeL.9Kʉ"ڇi1O\klG| +9Ț# I'#2~Io*V@|rZ1&uέUZten 0>}JV`̦lb'NiV5h9Ǩ Fj_XQ |rǞ W =܊(h @[%H6%6!l~ z|ڛf18׷hU٠QPIX-u¾La0&T*!oHF*)X+.=sRJฌ`9C,waoPEjLIݥgj$.ür R+upJhJ/`㯟ڎWU9BJ8#`:`lAZh)!hE:61jA&,c)$%."_2(!| )xU(B)t:>da˲Fd-?o>%덶)RsB(\j{?ƫyܣ*fi}GĐaR5'QFrGɈ"8 XS$"LBOceNWhk=Ss|g~gN2e}CFR(JBј *Ji THa &xD8JX Ϝ̐I~@re3,Ma[8($ڟMl!/QY|7 亲]\cyx۵+W_PLݵZr۹Zrj[/uKX*˜DcϒwҰqN$hB͹Ma lQvyt NII*j-P;D4o9  `[ZT7jRC$mT[A[悪L]Ph8b(CVhOֵ?+=,='M&3}O& u`y  c3,٭g5[<:w(JQreJ;ke4FzU8iY.8wjzWH.rxzhʅW(e=S^u#VO&.cZ [$h~c*y]F/yJJՆO_7n %SIϱ]D|g[ã'R{WsVNG1 0ʜ]80rԥKm9H!'Ygo&|pkg0В(< vl 8`oPv `ZQrk|Vi7_|2}/hzy? Ɠ"׫s ^Www!pkLVlЋz-(,=vh4᪩Gp2&T{x12>3JCdu` !ˣfIFZp:҃(စ2(SSl bu;(Evs8(RƼ⧏rPv< [!+z=mJdRO/ kGh-g1e&)!}{-8Y%λ%ݲԹbCPG`s3BKF: ?'鈓n$38}Ţ@jtDE},u;8u`颸vtIԃ62|&-Ƅら.V+f[cb7c5d';]T}j"C."PH9vfDH'tbP c-&kaRy,;Q 7Q!Y;3^/,Aw%Ld$zlR8O:1Eh~&r +(#)σQ%( PhqETE "G( HNvh޹d 8sQux{l" < EGCD3JBBcI pTD$ y}@E$cc'8T@@~"/`= )!d07 b97-&}<ӓ1oӳdV{2C? sÇ/Bz l-yyi䭟FOJ4mb)[>t6zW|nhpAe7[vLW}ZN&l3$){OWnBh; Wq<ȵ۷3SKF1(2րL|٤Kbsj7 h'(!93ŏڍzУCcFeKO;|4^^\0i??&ݼ 0S9>$:b¼̛/ ^Τ|_| -{(ߟA܋,`/-} V wgnYXb|m [aF #n6Nt, zǧ m[GVOVX?Z9#iM།@5Jt t@nN+!?WaZ?5\IT'eڈlt6v,@A[~ja= XHA\,@U.0`5I4j&􊊶墍WMVylNfhܕkH*3SMtng4lƌPސVpȽ ѹ'GR6;\w[Na18!fxQFbbtJŌbkF ڜ>Ԍ>&)S*&X0^`-Ms5hfӔiؐiU&|Ի ,!9_+z6,BC֢SpPypt  l9 kƚSh2B[H*i(Sx8pht=QZCufY!]4,DM܂iN k|}oC#ΝEAqLg*9OiE0s9$KIKdrh+ַ{R3(vgZ{rы魮V#yz5f>gtn&)\53E,@LP; Tki ~_"&5dMH^(TrӡKT-3~4c&Z4F1,/DnK\.1gT{RU2σ3T.RFC*ZFx& Vg8$mкFu>#ukrXgg5nUh\E sJi焒iކ#JF6SI%@"]ՙYMOD3 吘S dy0+?% Sbw;%onʊ{'.bɛhmL`iil+[9J]J '2 fUO)MBػNOb1z⊟i,WFGnXlfhk#ȚI+wОh^\6:9W_tn[hG=쏙]+ڎo@{#!  B hIN("(=T{T[_L._[:,I 7WAU=N 6X%#,&^X9",?j[$zLLt>{Loa//G6ȿX=:hn|,-;/ͼ2_F/x&sH8FHhZ8QgD`?)w. 1oȲqH| 32iwYZ6iA&TW;"5:E8̍I'}=/ziBi=_J. ) "%T?>Hy s5+RiZ)!)|| rCpOurKVdݛ\S>{4/K|ʹ\t!_Z>|MoLEsH4#Aٷ@}麯G|8(${\&{eh~fr2퍀{D$Xl>{bJ#L?XC_Jv۩C(sdfd3OƬGX5%iMhXmz BO=M'2Yn՛ebte{E`eY2 1kV͉<C M-0=3zP 1 ԎRS2$ɌuGeڗ\pۈVE:dRL&:ސ/\ю4[%Zz+xL4eH::5G.[>eyjvCu0ח&8%Ṁ\Z^&+ۻRcw"D!>ybyFͺKbiPA<&X&2톌AΎD0k]<<%^zΥ}`@vnbQ&q0[fsKobol5=sVMWY_ $`+CpsPOi0; b1z<鉲Nӂ3aCA2 j4PN!ըr^&dM0Qd,|:)9dž- GLHpxX+'3^?`-M# %^7~#]؄nOyī`ޑ\B陣zy0j>Bi>돼S8ZF _l@QA?Pz%m/'-aFyɢ&Q^Qcd32hUnpgEVIq0WZFPQ68NWȩ+jj[$M `6QN[jD2juF}ByGCyq|,l- IAa[7lMB!f!c \aGDz 0:DQ/}( 9GCle3G!aEQ'Ǧ#GJ wcx| cTRNZJ@ noB0P#o@}x˳.=;6EM!e4_E57j|UlQo#n6!U$~)5>+ae-*",(BeTwه Af<ڽ9s6F;OR bSTi)d+5Ĵ\ SQ"T޵7 Z`:{^p4zH OFےv46Ds/Gz&tez8[ qs3tL&H.2o]kO[?~{~0?ތ>j e^\?ռ_n'_ߟ<m[zfzB<8;N;IQ,iD)K)ɴHJpp>98|utѓ|A@B|k{əM~:ͧs[c}9|s C ܛImW=1 qdmл_ّ|~ə-ϼ~a~ec?b_M鋼m57)X)ݧo/Nį&9|}~ſΎ_e? 8u^zt?..\2K*W/}yyr|x4w~yvtu<~u%\s#{@y7-<>!\`dgOٳ/k-u{Pwz_oN6U3u4'g0^uLMǧݓ(F2**`wY3rҽo`H-3k+g_\=,J_ʿY$2$!Wgݨgo̚~u>CsBMO> Ax:胝Tpe7V"ϝYLQ(ռAQ*dvh~h~׽ҽ}뤛s+.kXgx)au.ʛ(=eӹQwq/D^w܅SaރONz3~O&EۏNɧan܊VM[$\6+93[@Er^oq^Q^~r4 ׁqM~o2Mxp=ܦpt=}}r>D9} ?_ 殅5_W!V8E@ˍ_\AuPnsQGL4rϒ?ټtg,-\&(6;s[ 33=M^~p5ut^w{ތp#)ͻ _Z7ӬtL\ҏ?_GmB(viX]9۝dRLWf(#XXp.d9z1{a=ThQ}|dGkgd""@8(/ ۋ Q6ӛ#sb1XwPƹZ*P\cA˼]YT  ݽ*eA.TgV7dLW~5.׳~ɣIϺL Y"^Jēus>׿ b;g%N؆j04FRA]OPN!; ڎ[띵zZg: Aҭ1n̚F"k6'\7K62DYl,-{6e6sbk%~0O28n󜢑qB"Ƃ0F1bq1*TRgs1Uz)Ωk#Cp5Lg:PYSDno oUØ6֜ƁP< M0xT42,DaF7˺ynl<4ER]g+]z$F7Ð Ԡ, &cS7)P49VKh]*nwӼ9n{w d~LVxԷ`NzoEaҝMzhW!ÜqZWXgkb55?ա$P]FVיaUk?k.1j(L]@ N%pU>NH-5U࣐u4.ά7nv%l .0-K9!#Lz;Sʎ:#DV`[X;rEK8*MqϻX~eд٭g})x:5ŧ@`KRuoz@`y=p瞏nۥ;pÝ,GP_(el#BFc,h~3l_gs:aSJ# Y`NrXΤK(;TTNtVeҕQ+ĥ±hv߲I9hg=qUʦ] ٍ))R1N YbxJyR0Tlq{Ns k;cDSE-]8ER aN$\V,q3%!S˪tΗC7U~0ӍDF@7}ZFS`ݘESX})@%JzS;TXaRE- *(IELwd}$_SדTKxEnQVAQ 5"4pet`zQ %z~Dzx:{w ЩM~ ۰B֖;G8FK,M" e@mSb+MVG$1[aՋVm)z$B>t?d䈪C/$"d8pU檆ldVs[YM.g97XRoh5|>c2nM:0GdcG{۟eurP0A+A4AQϮ},aåKp6\ Qw5ؕm3Y }Aƶ?>w:L.svI'1c;z0 쁁&uST\eصFvxqX\4cLBBzȵ+T*ZØzާs}2P/xf֜zfmU~_⠛'[OsOV,dr-֚hɟuJaHk+OJ9ި5xݚ398%sBIN6Ǥ5Ǖc* J !&pIw¹i L4 ǔBx&"4!*̪EXFh\NtG:*cRIߙ4a&Y,Q)KKpdH*T[!Egz#I]%5:F"6ȥ7\J *hHI"hv  ̹Ex2t`ȸD:h4rywj`vg&m`:ȧ$$ 8GyCgyG6vѽ\-Oɔs>k7; | ;` _-ꕆ; = @Þ8Mj_2Mj(X@&t|`X`Oh|/ăF(pɠHgfs7ˬTuL讬cDhK#BY.rVo\rsKhC#wҧ 톐]! !be<VuR\`(qGGGQҊjò/Ng+4b: lA4ӹp[ls[x^`r[-26N!ѻgم73 Q؛/;9g0W_6{L{[7`v,3딶/Pc&VƏ8+IW 5LQH43>d8F{mE#i|}ŤO7ÃnvKfmر:-7u?H&xPJ͞"&K/Lcs2ˇQEL67NcL(L7L"S hG[Grad]IKJ6BrW`Cq,ɢr4ג xah\Hۮ[ax"ܕocckpԌ@<ƉD !('w't$uB(3rt;`b#8gZH:XZT\1lA9JR2Td(D/$TPߦ!6z-etnvnarr9^:9R\thc_3V.33im&>,IC};ec.6A,k{j +b'[?lWjKS|1Z)(`k;E)8U+ʺJ# hS T%<[w–Z9Qtk'RDv25 +Yj[g7`rV쵨ʌ=7fyULLfmi?Պ"X6~ <iF:,9xIxkin %l%JTLIpϿ- -}kA3޶vۢh_Aп'S=$os{/ݞ&  4@8|b{BȞNt7VÉZ3BhHtcWRȗ"Ts)J1߽E?n@ώE'\ ͜j;VEa"/Q7ZiN.jZKX)pɁR^c:e,>!HHr*+]FͺX0&n+1)[PS*V/CT:kLI(D!vyd &ĢUJ)2'9Hy/I6I'/\2g^C mr$@% ĒA,3_,ʂM$E$xauQtFF;i}*, D E. ^y8_/5|*qX?*/vqӢ??Qaعr Z'Oxm R66>n:;L)a]xZϵ4(Ht:Kc]qS$IDt}wHm=HQ{>!Sm_0;EN.xe(k*>^ѾN%_VR5t]/>د,7z%o-5r7E46a;cN7Uńah/aN"YѳΖ|6}@Ec^?ޠ>1u"ֲ:1ccլ1S51,چQv뼥( [}  d <1C;5fh/Vt޺%o#sŽ0%# ˾nR*e/E(p;GpQLAz| &To(& "*άLyR)yM\-_^3(HU`02:VNѳ<-idfR=4D!S Q&_-o"%5YS2ي+,S*j!bV:eעA/z"%O{zv.*s%*NҬ1VWBKdBi0n> f̨2ώ1r=7_NjYau4`zW`V3AR:K0s Uh>`e3*oqnlK hUr(z V0F#[߂mKYȁjcŝ,1 ڛ{Z@]k=cZ/@47q%:+O?[EѭF1^oSǐ`=PSVDBVb׵ }]78; CȑjDG,j , 81 +>dy =Ƭ[1=fj?28A:mL&nMy%Da Y ֱ>sukb[FѻnͨKaVZo`j_L^;VK:6Pw( R>x'=9ӥ+oz'@i^ilQ#6^Y'őNk/z :{i :z{KBk^Y@ͭ[KCD*Di%Pd'#(yN:(B0ɗ[>ڷ8M2 k.+GqX"X"1+S$}r  ~-EĢ9,42@(wsjDc&f95aG(^DN(hn5M{f27d^]sJ-p7R4'be k7X)(kyQ(ܮ6ޥx91@0b%`Zټui%yђeeآ3V Ҳ@φI.B#=Xxu%E#\dU;K2e5A]̿$Ke36׹ޔ-'frv_-k)P rQK,V`hi!rg4i{A]JR=j QRG:>$)&Y|4#*(%2J޺X+T.?nO5aIvbP8!6ELP̫&A1S)Z|4dD!a v xIi%d_l˴F$^fyO[T婀[_ \nr.t[d>;|Mt{u!Y ki :uOGބxJͬWiA׌S)EIқ삗<C-dT )!PDfs Ҳ9Gb PEh&5aJv6;Lڗ >9ZRXRUc#ddYg5Ey\dŏC C84tgy-$ߨ-=}^UZml1od[];Djz{8g2*vcm ,n&c{ ^ݝ܉Q3`]uvnPMK/UͣjMNuf롖Ej5BaaJR5\k7`QKޯRrMݖ譔hXf$1we?@`є۷~oHmU^twxWa[Ve<\| NY 0pnUeqh8g;g 74bOWf},g1[wCh҃@Wczy6h:텑=>a[-l~w#@-}7Nh&2@pDm z"i'Ӌ|3zU`K9RɔO &r/m%Q$}?1ApPQz$[S8|c'kv L OK6nVW|nRgD_.O1#Q>jwdw+)Mpࣖ&Y,樥9{Q[!`?3}3n{Ԛ^K\\$Ȫnnݚ ނ{諃_-~c ln, ܸO}`l;ʡ0c CHөĥ|X_( > 3ބT_7P *qݵ{o?ݻNs+:h.|h_db~D`[fkS[WTh_he}zyä鋱dvNmf>K]o ,:wUp~;!"zV;+g'u/<a|]iAW:1n8v k8PltwK(Yx3KُonY8 UDi"Yti 5NFLpZB”6_4ĕeR9r+.VmG~=ˋ%ұZwWOd_/߂.]|j7ߌߔB>ŽiJ~cYc׊YԂ$]7|NEQF. LDHSN*6GF(bk,aTUd9*EFVMTCYeu3V ՊU$vj輫f.WC~`E6TCA[1GGO bli^Q/u䘵/DC cUB srt (SCZ%FqoET8Uzy6r*?ZqzT#ĕZG:EAELeKf0kEJv![|?V"pb~fMTK$9~c~AOYci ƕ0K}RwfHWӉ[ 'ap5 |9ލon~r yU,2tA ӄ^q?2 =R6S-6KO&}kЫs(oXdd=cWDjIC:%lֱ\9m@pU̴GC7W1V]⼀SOnLߍ_tV|\&ľ WΞ0z Jn_~r9Is p}cY>D!*W`闪|"'3?)qλAhj4AY.@˚K$eɗo,eu?L8V z|~8;gJt&x!G-3&"~8-N$K#1vbTPgAAq\H *Vd)*" 0GGc<ݕT.h5$Q$p;Ѵ6~""}ۦs4Y©Ă?uP6m` e*8Mt;N8QS IR#xֳWGͧuZTK`NQvg$cS6М` h>'(iQ:JbympzqLHjmˏ(O9$5gs-p&Dl$R6=ŢkmPS ^nE^BizU~J>Xˡm Qv 3e]9#At Y*B hU!wD= \ t(NLdǻR<[Ni=JJN VwYhmV[$ s/r ̓#$(8g4zecƄ@3j A@7c #j PQ06D-ITD-ZAY#c:*/9q\+/i&L(#v9-< V#Pѱ)wgq7YԪ8s`V2fhdMuBDp7܎RیB>b+BQ 4Z-%!4"\m=09zF2B">1 "A(:@Q5(@)4+K)ȲfҠ81 k(TL! .栴& MY*& ƃ3{O!εz@v4Zc(P5%VPQGtfC#JL}B37%z4 F}^>gė~lz.v[lCO:d0T> S0Uݎj$J3K$9Ҡ(F- YU\e/S >q6b{&{!1'kޖ\G6kr*]msv4A7HpWgלl3ft_]j_Ҥ<:z*I*s_a1 # j,]3GT9H)!1Ր\vf'MPV\-޽"8PPUpTf^BB[0߹/Ԥ~&-Bt˧~qf2օO^yEK2!io7<0+_4itB"@x=-Ćt6_J -.?^1wKT3%~hRQ _1g$5HAu{WO|R.*#M!E݀j+ЧqLn jx^`fii01ާrhj_"W+0rS+A4(M!XLۛ~d w{.dwTsBO䩡{[J:״e%O-_=5yRX; ;)@F@ %7ܶ@L 6Q'u锯ˌB 2l=uM+_u]qpp1p\~ P%1OC*F>@4PyjF*Kt}ߞUvF,퉶sFXT++ͨev}nPM:c5,f[QjeF]4FKPkjn<ն?$ExH/9A2IKLȹ]c*XW5D) @7L?ULljpgnl\eʓ~=8.,;_=z'+P9|+L%k:u07tاs7>QEZ|bTNC%)]Ž &V#{ߗ[xŋo\]<0`nesytAs[0+M P-yIar4@2}fM5K Wt!w].bR: ZTaE4bF:kr,,t`jVb9[KV;8 a DbWGgEx5誹9?W&jd4459nIFaڵiv}ib9Β=y<'$1br!mwH:ǗR-s\_7\;+ ۖŹfl.pwdz#*k,HAZodxN2$%ÒGrԼE%LqiOw?Y?ӈ&뉩%*4Fsr**{"? uB (.l Dov\C*ޢ+yzAt"5.S.9:2;eXdaIj zȴ)Lk2a5Ӫ  zYKQHilp5HĀZ g 40c4ZqY[@Y=r[ !Q,PM#ՌjQ4i@uJ>,T+{h^{.Mxjeq"h(K"(pL(&V0ވT0> so "Z@@ B *'kJ@iT`p<1Oi#yZ1q$A}G .*X6k2_B&y;Xgj衂 "6L0o1Zƒ!5: Be(wmK~ٓE(!;c,A;a.ՏY_y~@k~߁9ЅRF(U4|xsJX kKEL#j(`dFG!rnMLy;;oEծQC{F5秳uo=CXTj5Cl9*>z\=|ͷ5Qh6f9AA} ջF@ëwyoO{q3M~`Wk*/z=3.za:mPCl4tB9C5von}Ɏ&0/ؽv<0DP)pisK[|2K+O=߱BQ߉\M&xsѓһYKi6wg RN{'9%|zx_N`Շ=8/s)j1Z@T .85}rQ4lJ|\>^ުz_4,g[ ,bB:VE|1Q)~RIʉCW`6N$&8ʓ E<7Rwjjchۍ#h3`nb) !2MD f)Kij)4ƤfM|G)7fxᷧ5㿻x0 wo;W &`cD pkzZcS1(__OJ9̺hPS{\džB!!V Ӹ!`_ FւsQE_9*/"f !:&җGMB<҂ 5rVWH5zE`"rfș6ٶM)}fmٶ&SzFHs}D%@(t',o΀w]kqIi ECSsu+3n` no] 4˦g"#$8sR m_]c(z*׻{ٰ<4l𾃪$c#p,6PV7v__0b[wN|Q4L\{~ HO~f^|^mm滙pKGdW~X~٫evd~Y~9m U$jv(;/4}CLj3 {[7,VNĢBBK@u1 o0֌Q$g\F&NUk8ӧSDq&C2udåڑ@'쓞ei&jl^zHm #p$ ӒCs|V6UA ݭM֛*˄oIoLm)5%aRS2@A<Şp pKvP-d;R_mK (ys;đz)~ !{-5@g  k_c^I#48#vl.Q\t7 D4J311/ (ѱx M7];~A dW<bX^ z8]!F+!J)1kD[[߷ޱ= !_2̍~}I3F~xHXR  :-{{.^>7`$~[^mrh6BXh(V05J[׹h^,x%4 ۰ 8y;ޓ5W;5oL}%B;5HVt )L%)ϴJ1Wۀۖȍk-ō8y"wC6:}#?byᰤWے " |RPZqa?8O6nrjz'&4 }{5>d4ժSf#rB$ 6BHxDB}+¥@e J,7VaFiX{* c4L݉*n_EQa`f5P )e"E|6h"=.hE^_ɻ<}fOO(Gb_N/=9^|I$սꦁzn3*7|WXY6?T@F7nEts˷za]D0#CU]g?g^O"^{XԠZϨK#O[k~*#5.ŐĂU̫Eyap+<ivJ0 >iě@wDA]*4TCu mJ6@DZԤ)CpO&buVܷ;iؔ`vtaSV{o.@S-ڲSS o@\dL=&. p!!!FBuy4]b\\3dΉ8qPo[\}% \yo JIW؜`k}WԨ9룠n3_5ɶG9:iaJq)tC6ctctݲveP4&NU&mY2P)!c!b+HB̀.M K5e$Fxe=Bo]ڐ&dq57.e)89M8 NS S)T !S`ThhCC6J"6)_I$bꬋX))AMؐ3`KVS'{8(BSM~2(S[ ʾe o3,o|& @Zb-VTNd *Ws3UTc84^x'񉦊+9 3yQqz`8i.üK[j[j- T@ ^ z^zHm5g/=m/+|[65a Ioym/sW-hB ڎkHr$7icRU.= 緝*Ѐj€#`m25@ qAX~jq&GMZm1+>>^fc4c4d|l`hf2B ?YO/‚%UBál`-@T5.L(vD;S ? VhKC;ڒ:@R@lo_[-"blwE#=ՂYRGsC~ f iik: 2WI(:0;Xg1d8c]D:]ZyNCpWf:5#j%7ש \3ՊSLhN@=RQSl`kXeم TX}\s}V.lk ( /_l?kcy]mnjھ.JP$蛺̾/"\ 퇕HJSe!STIXHPK#c %}ԴUMUʍ:Q=zI 6ϨX8o}"Fgna\Sb1Y!"4Ĥ4l7eJ(fNP p*7T \*ٟ½2@qF uk<6~\A+eJSZ3PPjp:4u?U/L2M/&,e8&$ 3-o^AitB)`ӿS 1Qz#pMcfqD#?bo3!q:~pNQG'1iKqO9ˮuBb: T_G4s`AZFjBi|~~wW;R&;o'| q^*'P;PVcqu5:WdϷ_3)L`&p^v9njaVOj;ڑ>||^j"8,pfQ;HhF$rX@oF?D%sggmZ&Y I "14K(ʧo e}ug $JX kKEL#j(jbjxrnML}}kS5;Z{7[,0KsKG`Ča,n"b>̥O˟ԑ0{Yqڑ܁WPc|B,Mnd*TH}Ez K "3N&i8gs,pej"5FԷ 5 tG("pF2>"pG/1OyYDS|ZŃiGzHm53 KAAζn@t^~sw&`ru~6Ǚd@:PFQdլؽIDΫV{Edvnx?޼ƹO>dj$T cP-?;'N;c'BiN؅֥̈́#282 " oM,Mj":$S4KpeY^͝Qmtid&@ M8i i1vV'N!C8 VGt1ksQ9ݤ W!Bݗv dbgτ:Q(Mˮb]C4;«hɓ{JYf}Jn5{f uӤRjD/dޚw<A&`I)8yT!e5([˲6+xZeD?wk|7iؓyRZ2ZϤ$H06R_$e{ӱoe~^ϹdgIIM`!Km3/f!Vc d!q?V& ^Cڠ>E] ]HH ]H5 ]';6hUe=,A:mfUbivQ1AךB>Ww;:]H$uɯ VYYyz;a]^01[KIw5}$\[Z۹T iR R`Gx(ZQ ka W̧`v׆y/6)264;c5yLRE)nz@Vn IJ&'IF雽\Ҿ4&0EOۭC_[0|]ێk%Sg]ۯV~ }-sΨGXz'J4rmEoVtT~%5v~%ߢEjcر[09"9y.6@4RV RUkM~_HS?зjLO0fz$HBL}_g Jcǿ%Agec5~7zIlNxWYF\lP;np0;iQ ƲACUK-l|a!B`S,C"(6.Q_c׭ttc;1(keK%R jv5K^$ ֛|;Z.D/yzcv(H*|f{pK܅իyLjD= V,G1˖BI9R27&'uLI_DALH{[2ꖱO b cI0 t0qHclj ntmI.'iΝ0hcQ"9/5 Ѣ:m cNU '_bbSkS@G%W ӉG 3vo0HO޹>p02^خ0O.>9),*&|f(&֞v= AJkIsZYcZ36os> B͆rv90XL5c#R 5 0eK%:ĠZ4r{o0%uɱ/6N:ʌ9&3- d8\C)0pZDRiYGا!ZD7)|!HV@Z@dg>mtJ.1.XFX?#;5گh7)Kh^HZ}~ɬSW+_~Pu &7E9BM X.Z̹|S `c+>={Uoo.Ws'8ڐ#X&GaWʜiR2H3'9nRom{yqrlk_AZͣ(PLFҴ 2 2=pWv+qB ]ds_GR?aq`'G嚝Spv:)xod\-!hAJ:U{GMDc~S!fzn&,~]ov>Ζ۾ W9&}Z\ }8} AZ=F32fAP馶 $)b|='PG)9zx+,V∳{FMmp"tR#aj &Bd)%Gƶ2S=qmp!ǑeK=dZI鴥S{uÖ*MJYm{)J)nR ջ1NfY“NZJ5tR]{gBJ5tY[|jJ>e)y߮}Ydl鴥TnR*E>̬_~ʬ&IX!bJM 9L1+bY^$hئ$I/m1M>b$E| - )7K׃DY?/wW(-7 rӫj/?4韮>ijR4cj`\l'mnIg=چlVH-ȑ8,!+߻n-u$bTϻ;ÝXt3(#zߤsDdVq=-~6r¹U|2Xݹv|[NF8/o*պ`oL9n5ri DžkYؗ@v`r ??,ʵ]^&*MM_ (bUX _s'j ޓމJGM? 84"&`ӾP^ BS=lyD8m| ;RZ*0!G@þ4thc:Q9- >$ /T#W#S:LDCr)I+,KHP*g1[[]q[y`l%Gg+3f0/Tّ4?4@ ZN " !G"rH{˛`ї*0ٱqqCkxizooUozyZ)||f$Q(:}īJ?e~gJWc8/+"o:d/9 =N/טC< ;d''Q=LOn5xtSB?tyƒQI)Aw7QX9i;NŮƵזN(!n<]|T㗶ёt-}1&_ni;GF=8k(O8G5ZrK[np#Bc.su]lynMw0D;Oy@0s5үW_ I>Jn˛3'5j>^g} Uֺd:v*t|c~Qc½>ƀZ8zS~V&cٮ-vs_1߾=h*ڌvƗ 3Vb%^߹[DNJ[=ȰNeT2[ܝA[WZ1/sИ[KjYjn;$@_T(F Zu\;.Uqd4Udzukt7=Cmz:z˓Hߍ'unuKk%ٴpG-Wܺu+qʢZע@+@˹̷k݊yCb݉T?.n6UVң~1Z鬸K z~O9©V FZXSC Z$n-=+#Ww.BIÚY(qM#5f%l0j ;Ԍc<1?l{@׿v/0} @*K_{_2{"5%Olν0un;=yصTeXENW`<֑9:_<6u~|?9>;|G&Хyx|tHf~6eZnb3Uu%\9ӿ3yzg4 |P˔ bM!ZS3u޻AL-R\2ޭK۔YW5[!)zq>n|z7_@e[Hj?_w!_9D1E(0?I&Z.bNnU`г]Wˏ]Ñ9;*ƶ 6yx{Rpژؽ;yFʧR`(Hk N 1)PZnl1sPGKEc^h;33E%prM]NO[vyn 쳘Ur#(*gZd7iTU!_?] o1KpO=tvNN[J=,g- V,ΩY֭X,F-,Mv";\&=i9_c3활8o$9FR,Ss8 xS-2~QZR+@*7yCj ; ɒbB*ys Ujp31T1kϗ }u!Z>Їj_yuwӂ)j|z7y/RޭPAe{͛ջՀBGO"1]։x) D:\E b7^'M!Z'RBr01#V{^\C'/ΛvSDd- s#jبZvԒ]Gme?L@ń,W̲VE]m{~r7e}Jnl7Q7D| Xu7lv{Nk)W*ٮO;O&}Cs&+`PeLMIFq"'QL y02W4X>+|3M$<'x|yR2$ o1di*e!.A{__?WvKt&/"sxh %ڰ D?WcsM]8ƺ@kt8d~sIh~7+&0kAe$SwC$]ѫ7jP.MfNԒ=cJjFi OMooF>$~| L"@磤VH($GI Y3|NDž,GI=S)=v}̖5 n=+Y._r[e=N<nI ύA( -sF7d)h&XCf<`sfb>R\Ң L:߼[VF`@&lj2J7$ȅ]ob@NXj>5.g& 0%%0 D8 1BE*B!%1`|& J%v<,uROb _bt$BP=bC1&0$(da(e ~[[VtS\EM^2xC0bU>|) g-@ϏYX,F-,J8gVBUޤܘ?ҚΒjDؘr$X{ W+V\*|a:;Ye||'#SdV{`X.ߟlw*ɫNpt3SOE2 ('$l݉Ky g*]PXLe"Y CcfI%#PXf9[0T龜k.^|@]ӳ,JǽW'Yf( c*a`$@ J2tE\%:Ҋ$0T$\D4qw}bb7c=p@ߍ>ooMxaavc3IW K0 ȝnBis[Rk+z8v[N:9+@`mfC"4H7D2 PJ!p8Aa P,N@mFR-rװ* )E0 ƈ8AA +\#nFG細BX9PZ2U'T {e]>Hh1xWv=0s(}XUX+$z w5C N|P 0ą)U,֦a$M|` vsF37J~td#x RѠRaTmՌ cm8i6Ch8oSj0tOm7Se7 lB1]HGL BtH?%Jj]c9B&1၏_n٤6=3]K0JA^L׷s-I]iӄ777764wV 41$ "J cB%@A@UK }%޼MMp ̵)ᙝ.G 4  h`Q + t1 hJϔǑb*&)J(\%49PP":vVZX4s#}x㤻g/E^:d,uw,>Q#rT#2]FpP$q(0 aU(B2)h,=0v!2/c4RUH"n( `*D/A1DZQ>jB? ;=CQvO .3B_)i{k=ңbRQWhjR1=|>Cc4KQ|RGLjb'v,-F~y5NEn;٦IN1BIt8I=AJrVLǣ`aN<L9Գryl*?~쟷~X\HUkaEhSoC 5ǵJJ*;عeBVX+GU) s. Z OHlkԟt zT%D3;/KK GM'N*IsA(i1fh) w6{kX\JQ=aճjk30P 0R7=T3'q.(cӃczpj|էzC)kq@$uT,|Y֭X,F-|,JŰg/美?ڵnżHG^@^HIZqW>Rs*_st1k;*ÎSV!AobjT ֍u'iN9B)*,3R UjTJjjGy َL"qyB!w$};jَ#NI%k34E/-Hj /d;8v#K*Hw:]DN͟xL@5өyi[mɭj^*W,jd9FOR46%-v,ߏ }~5_Gy'>NO2w`ǝ`^ ℹoWy왭i}qnTiez+Q*yzU;ʮ91Y;lP2`wSRRs$7JtC)ސ^K7R рҳF)%n(< CޏԳfPz~(ԴfԳH9Q J0q)F(=%,/5EsD[ɫLK^!D)n(M6sҳF)An(%(E)D@)gy9/=oRRDហr7Rs7c@١KB=3(EkgT(=o"R#(=%b?s:R_Z.QJS9Ny07HŽ@)an(MgR*PJ(zR*PJ-vD JE/fRRsn (=G:D5>JrR ;NPz(% $ۉ(% !.=oRRjt4H/PJJScQ.S\je}/=-,/ (=kBRH^Գ'RK47JpC)J3e.QJS ҳFSEInԩd/5bK1.([^:ƥC;ď\x.E:q-wˏ; KEE-6 a`iݚ6R#E&mc1{7 :N= ]{htzڌzaqlw։f]]ecl=݁Qw$sW&_cbHg::h-wqmzhk;4u& jYZu_\I] &iԶ(x9˯uV(g =0IbZFcSɁ(FL%{cf99Mu195jwaL7BfJUn,mU?޽$@ ֏!o^ǜ RUy4~r&>_D`+ϯNSYެ֌p[hҽ o>p&iM 2Y}vT (>;8"-`x{ue:>ˤi>(qJdŕwz;x_ L yBϐKB\6xfW't8郫|YNh ƕ섀z8{'.L/kR"'~QmlH0˾NZ,kE5Hk]•í~j!1)XZ)")1Y•u`2%7;aϝ]}ߎmo@zv铄\ n0o(V9:@5٤׋/W͇CەYV芞gFs=UwWW߮:pyU{z|ͬz/_[d[pmq[]x5?=88=*2J9-?[!@MP )nO1l+M}LEHՊ|Mc);m?7݀'`jKc6k ֭&HZsrʅP๸]ٕ Ɩ~BGOO\E3x+ kmd?O~zj ZLcfV,gs;!כkX8b|>b[Xo^^斀 /j;X ޟm|zv]C7 sSRnVYYNlj3߷CR\W7gﹶwukGZ^٧[\GAvQ]Ò#̪Feuّp)L~0~#Ļy>Oօpi;ލZ؈-ӉFw1&~oӻua!/D[GUxt:Ȟn^G,wn|5.pեI輽-?]3aw>.O//O^{#oo!~'0,H cOs=!>n߆<@6Jx^!dABChXzD a-g,QJB~c;駻ZL{df2}g4u}.GJjo>g1 hJc_h_ vǔR,ƞ[*)<e 56_=i?7|v >*0s[$pt_ΈAx(NApbTlp+/W N-phzq~MRQ0K2rStLZ`d@0z&,IBjZo,67 F H{!k`P 9U 1n"bğ` vݛAK)`O=^6i(@ T;0;յܶe+L[ GE''o+͈F i,x?j07cPLblV=, $d6~!74 ؀5>8 ײG]ar![a^޳!NPD0xV 7/l?EQ-'Xכ{՘:1~5uf-!ЕVxt"^486NMm2KNMQsþs0gv+6z*{6! m]3*2 bK6r0 2M 6E."Axr yz\\ 6n.V_k1JUnB}GV7Fb.%<6%2cH>&)m:K IYE!iPRx-gf@RjI@ x{|ҾXm[ܝL[% i@/{Ë|Y.6ICj 47dL|.g/<*qЃyY!=R?̝9{12EԆRDLGC%ǁSSR{cB’~Ԏ3} 7.&2m^AYa藂8\!H)Gҙg{YBF5)>EjLD!(ׅ2(OKRJp["Y'^Pl^v|ϡg 4WHA Dz"M-\g ӌD0]HR 3,BRX7 V;=a )cYRd)˙J)LQB)εı0Ijd4- XRg6#yF!;\2ΔS" j%c0-R8gdPxQ08QR;aĞs9vV/Կz$ $B~s4RW\ʡ&9SsDϕ@%BPædJʉ$u |GN)kXvO|r )[%reiܽ}T$ 1`Tt,FLՇJ]r7žS/29Nb%"=9h OE$R+ʑDŒ|!D/]0x.Bg` ]o+brio5R-#G[-!0G ~ Q?Vlp ӵ)L \V BAX€[rXH.+-.jzjﴽ岪ge1p=[)L$u{j Cxg2uQ6/Cϊl}r/Pw,'k$S:s0hwgW Z>w~k6uFrס3|anqj2Ku7[!w[ˀpo|iVKR٧ue=U~3ՎgGM4ʦ|ݠ?wtb nޭ y&eSm7\qz7U;ѻbc:ncF!5wn]X 7gE +ҋp ˇzX"4 )/=VS? bDJD?j핞핉`ӍO?$-FiY)q(J1eAƬ}VZVyMrX >AƼ&4 yrQyxɎ- >d\/iy4Q !M]<ժh;l!0~rHaDIXVn5hOJ):VZ>`Q// K3_ri%efN3f8xo{R]wr<k܅]JuƢ`ͅpRN4R66- II9&)]tAFJtሑ] ֤d":FF}J̢ r)7׷NJ8yUNߧU"s:stZ)tDڦ~@G~2ziP'NI?Qc ŒKWK^f*|Iֵؐi$ֻvW>F1܂e/"o&ɛқza?^icXG`WCe=vZpDz4u3 ]8=q5=G 7qwU㢡+8NOm. kOr{v'cS|rB0sUqL$J.5Si8`D\Wp `(tXH ,/@- b<`w6 G4Ujnr*0N+ɹPFA*Rav%ؒ>oݳBr0f{9<|ʏȇ^N_@(:|7EAbۯǰZoEVczxi1-P8޵a[m>58m?D[dלG[Bl467 0F0Ԓ j\ 7]#"^dIW0\(gEw 8`xG! QӚ+&wmH4;wȪX&=` ec6\mZ-:H|lY]XdU_Չ^f7[Ϫ '(| L\~鷻h #p9rBnHF ɲr4*P!< F!7cr27sŭ!)5*򀧟t P Jj1=QEZ&&KBΊA2gb9e$-Y˶;L92g9&uDpշ)oַ7}).re:e$SYB7hkK}1qt}8jI(Mc iguwI놼e-ĕf/9%%! H.ŦuJa;cϻF,FȼXpsۀwb"Y3;zi/{gs7iZN^oz̽ kaIBm#kn8`'?{E]N)Rix'd(t JHu <9[W:0U3#eba[m~r,?~r[y#n~VmO4UgzV*|o_u sM|~_>Տq??\_u_:jcono{;νrms]/=|FF*t1MmՑ\D9kO͉0!%^7ξ l&Or}z΄X -/f%7/ݨwK%Fᕞ`U D9;+'qap%Q x}ik;Ŭ9^Tݧ"]}2l cĴX._Gѷġ^Ҋ1&%<hA+WtG,*7!<76®\ؖF7W25KMRׂ&様ُig|RVۋ`xsy;@y O9!]GbWqΖۊ05H{St~[a:N h֒|(bԒ$1:C^jF o/~zcʭ:]{ׇf։5xdUV9Q?;M'Ç>~^O/#~S$HwwݻOnbtҽ+ТnѐZ1N^î"Ut >W)YxyI\"&+6)M2!s_?H8/f׳_;Mnjh!cuU3sWwxmot^[6^UW yuSz/6pz,VOLe1sxXelXXEusr!1m{ v|-V>ЉvƐy-ޭ |p`JjCMx7`F :ΘSzn~=[m)b[mj,f_h \w״H`吉ow  %~2%H*k5;n']xkӏ5y޾يSh8[7Rk'61X& 1*>6R/R Q P Jj%p4nuGhY9'JX?KmP:jC) JZv m rsE%9u9fH*g()ĝ>̱̋<Ţ)86>JupV+|>?U\[nm1BeST.|WXer&7iu}OՃj_~_Ϲl~Vl*U{Wd/nEtXs\(Y֒ ]S~[ RAYg-͋E{߼btZk>Lm T~@57F?y|??ݬeOAmgO¼`NM|խn[̶xO]uC5 qI^{f{R2{.4_FvGt9Y~9cL'c\<c\_C^nm|AdO^ڽTӥ,\|xLr$0</X"9o*v/k.S#{" ,Ƒ 8)bGl"L3ma8х994쯸t~Z:Cus@̉iݧ:&5'w.z ҍfF#uF922IE*4%R@, ALD$J 2PY"O)3I!Pɶk^{)ZѡDw깹I$jF(%M`(T,iɅfe%Ȅ*EP hѯ){nN܈Zg# 5&M(YdFAJ9<+4II(Be"Їgc w0لxL2>fcArh2u<tQ%L+3KD(N +I%\k;15JHƗ~Xҍ5`]ōK*t s8!έd'~t0( #\i;"ɤK_Rh݇tB9GJ10!t/ԻN$S'2=G{”jhOG{3r*ZZ o}Ɵ` 1w??G ҇E_hMuI 3$" a^ދf^W=ssCjabo$ɞx[ls&~l%8< ܸZ̹nrUk.$Lq~+Ն(pì!n{M~w+ rBͿHx08MDCnB(eB4h6 A4a9g=C0;(u]Q ԡf@]t%Eu飤#NdL5R .X W剘N!-嬌>^j}bGj ՁngBq}6y< @hJݐAxXxqG~Fv_֦?җ|pFaw͆Cn:NgxsF({n<<[m)wc~c>o˼=qQRtsut۳ϡS\}pE כOH wڎ~J|.gl{t88k/n~wSvծbJaW!4@:+km D=P8jJDa9TPF7.v.í4)5'zNi #tFr@VU7Vq\繑)D'6 a"-  C6i➛\0)?dU+-;e~.-tєvhrNDhc{{®Iu.g 6J-(1cc}lݡF53!Z5+FܮN#CӁŤ4hOC57DFT$ॴۓK*_.2\qMD!)گ-"Q5zA&'dMy8>co24D/=8RFáX>jADV=9,Q$L$~P =ZۊOC<m93g`j8k/,HV?r鵼n|0SO0%47"|yJyV c2M"-R‚ð"ϲ2OUER֬'!0CIK)1J/mlDI, (U*%ӌd@ȒP`B ix!KE.}8'ΕIp+IKԾ ,V^jϚF VKNýܽˆvx>Zje_ácz+U[ca~$BMPy2a48:QpR譓ſ@qeMpгuu/7_M_PE5~ܯf_>&OWMYVeM]ʙ̖ϸsl>Eq? ~8=V~Ip(+;e) (`]<R`-n GT͕ PzLESjK_JLjRfPj0A1Mi;nC)kob(=&);t(+ïP=mJKhH  cD_?kD_Zcӄ1JA8҃@1MSADOǥ^4T4t(RhcR/vV4cF)^1}{ڗjJKhJFiC)#6rҵZ F)2?QhPj JLjR\Zgښƍ_Kl24\'[I*]{n@\kwt4TȑĹI Hepk6@{VGRjpR¬Tr$T0+-F*O#1[)LeaBOaZmzq[)aVJ_K5 +*t8\\eS\@ d X QmX0nz}췀w ۧUճ|UMx~U>K\K9=p,.QUNjK]޼;9y}3.[*M9J$ ~ qo3"E_sa^;'}-C*vn W8Z)9]~uunw;O$S.S P 2oFwM* t Lt6-&tP(;?&p_;^X._IUG |R;e ;_'g; 7aSG)tbPާW"4PLH]6+sP&/>sެXS-~~ P}{}_͜rlGdůSe:jr: {E(F;孡 KKi( _ϑjs)cSs4l~KOY[y1܄cR6wsW0niƫ9h.47am3*';mi;9Ŏ8 }e<`g4IJN8!=]-=^`xn7vXP+>O). =ueLj$ Ԏ<$9Lrj5Pbs+1 J̹pNJ U%b A9 `D43P*dRX[ͭĆ* /[Mm`[pfAeN(2SFrQ2F`˴mV2SE8?p[zn } !oٻMoBblqTV6EOT{&P´力+mN-&"*$-'}Lˉv[rB9;fC"i)]xN6A|9fT1 % H*-'ƣh %g՜{~̰2O@ ڑNMfzy&LyqCɏ~\CjiogpcGgF60p _]t2XzqU}~@?ͪ;"YܕlQ@V&W~1bnk>ok8u-o$B3>@_L8FfWn=z&~>9HUJ7Ѧ 5FZ#t:M8 8颦,{Tv[ 1eKhK-QEOZ!m[Z?uk9G"o"Bw7oKDme []u,e1.ZX丽o}Bw\/kejjcZ= _{#n졂GA3*ɽBok{W~BdT͆D}[Tͪ˱* +L:Wq:#.e*G# y&cSH`#wP;9w tbĻ}yNjWz.,䕛hMQ'z7Q[eѻbb:hƛyN̻ n]X+76%%jR89N f# VT`6RŘ,gEa/ܯy6Q)B C]C =jFk)R_}\]s?=3R\c)\=W:P2W_oSD y qQ@7;s唫35~-|u3R™τox9wgo7o<.Ao>P^nOHSNN9WIٜ* 2>)Cp@yT_2Lb)wv{ TY)o$)Ϸz@_ʸEARePUΙJ? $muNAksg\ZS0Sѝq& : ĻRsO[}nq RΗ $)ϷP'+=>+uuJ5BRK/=n+ R(}lV *Pjʈ]hlAiK 30E ,ϑ"Rj2Z bsH]DdLIBnSPu Gui4bv,U=VGݬ>UQ  ^-[4'߳"opg1Agi?_˪$>d~.g׫˲~X:쒢5vIJ{}CՎN܋%#9X]jCƢdb%8JګZߒzSQOQV׷00HLIk٢GVO=yw;Ex~klB( Υʼ\\ΥdV[YFvؽԀзy jQR,=Z2P9 "`cHTZԊrY֖ !Y׽NtV`vJo'ZMzF͇{:4tB-jhPLfZwغ]e5a9uK2&mwNՀj4l/r@F|뷵&۪gqokƨ{"[b/F/l!w[){F"OS\f?uq2.SM`U+㤽9 |;WlA.nzY-ߝŽQ?}Y.o޾y*Mm302m^Q͖pRbu]ib^E ުӧR;wgPr;ko~oVe3gCPSB .u׋ > |2 rQyxQz2`o2yٓj@D!6B4'(F%)qo ZfpʭuQmkmNNK+wbuWkEMgd '7OY,>z"~^c"^bh8>oͅrFpLNQCOq؟SQ}3;MYK* pd9)FLZBQd!N:-#AhI]^}=* b1[а ~y]lN \g" p;_̲Ԩ~wo$5>#Xu6#L6qV=J MItLR(KttZhИUKUp.ƚ6<0W@*D?;a}EI7k@cxPso覟R KeǪV ϖzPJqmh9XhAD#/u ?Dсg7QBi Nj"}^L ;956SV*? =8y@6P4oNe2!n7vk;C A`aSc{S.1`de14-$"gJFfi bC46wʭtfl .ѬxM° feZ0LŔ J"%E&d4@cgF (PF'~n z?2aתBn}wCqa?maVk_z}iiݖ:PL pM/0G)Ɣ3ť X+9+rJ :(&T?aHiBP rъt L܆x pv[c?Cn IwAgrCj$ebP B1ijmK |䛢ԀS7㧤b^`)yՏaL'Th.\&{^6'scXABMm?pӹ_k} 0^r"4V"@]\FZqRoƗ".(N@ZD'r2+ZPf{uq܌.eh-ꜷ/& F{= uLrCl?/@TvD=!Jສ{@_6:O-]`F]/yYPReIgX ر>ӲkJ14*8JtoOjJx`oimup4x-Y! ps}n̽pRjӼKyw\8 Ȇmͨlэ :{PpyWjlv n]rrWqF7pUp}A{JtrrP!hkRSpsS0pL<%>Aۖ֏7UAWA$9Y`*UA;!^Oi}#)eEGTZfEq m|9exTpwH+˔Y6187}6OVcO:|G<}Z2g3)6%v7{ '"4ayj>|{<g̕rF+?S[ h.pR ]V{4KIs$lՠ\g^Xm6 GGZNe&JG^Tt]gGl} D)UG҉R JQX5TڤJQXqSj 13805ms (11:40:44.249) Jan 22 11:40:44 crc kubenswrapper[4874]: Trace[328022024]: [13.805551961s] [13.805551961s] END Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.249960 4874 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.250285 4874 trace.go:236] Trace[2096886872]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 11:40:30.159) (total time: 14090ms): Jan 22 11:40:44 crc kubenswrapper[4874]: Trace[2096886872]: ---"Objects listed" error: 14090ms (11:40:44.250) Jan 22 11:40:44 crc kubenswrapper[4874]: Trace[2096886872]: [14.090831046s] [14.090831046s] END Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.250325 4874 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.251215 4874 trace.go:236] Trace[592521690]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 11:40:30.833) (total time: 13417ms): Jan 22 11:40:44 crc kubenswrapper[4874]: Trace[592521690]: ---"Objects listed" error: 13417ms (11:40:44.251) Jan 22 11:40:44 crc kubenswrapper[4874]: Trace[592521690]: [13.41725261s] [13.41725261s] END Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.251249 4874 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.251293 4874 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.251349 4874 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.256947 4874 trace.go:236] Trace[646061860]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 11:40:29.809) (total time: 14447ms): Jan 22 11:40:44 crc kubenswrapper[4874]: Trace[646061860]: ---"Objects listed" error: 14443ms (11:40:44.253) Jan 22 11:40:44 crc kubenswrapper[4874]: Trace[646061860]: [14.447552629s] [14.447552629s] END Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.257008 4874 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.266363 4874 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.290654 4874 csr.go:261] certificate signing request csr-crh9q is approved, waiting to be issued Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.304158 4874 csr.go:257] certificate signing request csr-crh9q is issued Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.459083 4874 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.459138 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.525119 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.526016 4874 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.526062 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.535104 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.658531 4874 apiserver.go:52] "Watching apiserver" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.660761 4874 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.661076 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-prbck","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.661410 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.661558 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.661678 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.661719 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.661836 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.661869 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.662389 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.662489 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.662542 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-prbck" Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.662613 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.663096 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.663673 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.663972 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.664364 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.664836 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.664890 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.665111 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.665183 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.665271 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.665322 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.665617 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.666092 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.669661 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 07:54:02.467067046 +0000 UTC Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.685383 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.694292 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.705167 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.713894 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.729825 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.737300 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.748247 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.760183 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.761711 4874 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.770876 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.812560 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.814628 4874 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782" exitCode=255 Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.814677 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782"} Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.821445 4874 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.821898 4874 scope.go:117] "RemoveContainer" containerID="ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.828335 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.837551 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.851645 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.855955 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.855989 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856346 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856385 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856459 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856477 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856497 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856511 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856528 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856571 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856586 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856601 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856618 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856639 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856660 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856682 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856701 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856718 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856734 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856752 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856767 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856778 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856783 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856822 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856843 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856859 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856875 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856863 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856892 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856909 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856926 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856941 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856957 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.856973 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857002 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857018 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857034 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857056 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857071 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857087 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857105 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857119 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857139 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857155 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857171 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857185 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857201 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857215 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857229 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857266 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857280 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857303 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857318 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857333 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857348 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857363 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857377 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857405 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857426 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857438 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857448 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857377 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857522 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857571 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857588 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857605 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857650 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857672 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857689 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857705 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857725 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857747 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857770 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857842 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857860 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857877 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857893 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857909 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857926 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857943 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857961 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857977 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857995 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858716 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858737 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858755 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858772 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858788 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858805 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858826 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858856 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858881 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858899 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858920 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858936 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858952 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858968 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858983 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858998 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859014 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859035 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859052 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859068 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859084 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859098 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859115 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859129 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859145 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859161 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859178 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859195 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859211 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859227 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859243 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859259 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859275 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859337 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859355 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859371 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859389 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859419 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859435 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859458 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859483 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859499 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859516 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859537 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859555 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859571 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859586 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859600 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859617 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859631 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859647 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859690 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859708 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859747 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859764 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859782 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859803 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859821 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859839 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859855 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859870 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859889 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859906 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859921 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859939 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859962 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859987 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857815 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.857908 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858197 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.862330 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.862388 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.862435 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.862456 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.862481 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.860918 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.862307 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858328 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858744 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858900 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858931 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859219 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859927 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.859990 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.860001 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.860113 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.860485 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.860860 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.858266 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.862609 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.862881 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.863001 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.865491 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.863105 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.863333 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.862760 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.863425 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.863574 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.863728 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.863751 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.864047 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.864278 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.865102 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.865248 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.865390 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.865527 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.865592 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.865819 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.865851 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.865858 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.865951 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.865987 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.866058 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.866145 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.866173 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.866306 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.866411 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.866512 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.866807 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.866811 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.866885 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.867086 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.867324 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.867366 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.867653 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.867889 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.868020 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.867438 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.868166 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.868432 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.868450 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.868450 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.868488 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.868724 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.868182 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.868781 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.868798 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.868834 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.869034 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.869600 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.869776 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.870069 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.870143 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.870207 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.870239 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.870460 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.870986 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.871601 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.871720 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.871772 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.872062 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.872232 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.872324 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.870718 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.872468 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.872816 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.872942 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.873242 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.871784 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.871152 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.873312 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.873453 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.873511 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.871003 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.873739 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.874020 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:40:45.373987573 +0000 UTC m=+19.219058663 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.874089 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.874121 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.874956 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.875177 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.875302 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.875437 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.875549 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.875628 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.875722 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.875812 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.875902 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876035 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876082 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876120 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876154 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876185 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876219 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876280 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876669 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876766 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876799 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876831 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.875767 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.875989 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876858 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.875726 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876496 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876888 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876917 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876996 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.877030 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.877060 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.877094 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.877122 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876885 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876936 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876644 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876631 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876591 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876483 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.877000 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.877031 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.876577 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.860764 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.877979 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.878010 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.878161 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.878265 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.878536 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.878884 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.878884 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.879328 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.879418 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.879769 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.880025 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.882065 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.882529 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.884116 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.884484 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.884907 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.884952 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.884981 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885011 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885032 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885054 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885074 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885096 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885120 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885142 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885166 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885190 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885214 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885236 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885291 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885295 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885325 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f5ca785e-1db4-4e08-9ad0-66158728b48a-hosts-file\") pod \"node-resolver-prbck\" (UID: \"f5ca785e-1db4-4e08-9ad0-66158728b48a\") " pod="openshift-dns/node-resolver-prbck" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885361 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7628z\" (UniqueName: \"kubernetes.io/projected/f5ca785e-1db4-4e08-9ad0-66158728b48a-kube-api-access-7628z\") pod \"node-resolver-prbck\" (UID: \"f5ca785e-1db4-4e08-9ad0-66158728b48a\") " pod="openshift-dns/node-resolver-prbck" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885410 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885436 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885469 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885486 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885493 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885577 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885582 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885607 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885619 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885630 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885719 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885765 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885781 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885802 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885835 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885864 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885918 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885954 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.885992 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.886013 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.886030 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.886638 4874 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.886221 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.886373 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.886491 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.886490 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.886659 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.886544 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.887377 4874 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.877171 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.887291 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.887615 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:45.38758915 +0000 UTC m=+19.232660260 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.887958 4874 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.888107 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:45.388089406 +0000 UTC m=+19.233160516 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.888530 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.888579 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.888612 4874 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.889101 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.889141 4874 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.889161 4874 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.889181 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.889253 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.889273 4874 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.889673 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.889054 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.889976 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890174 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890610 4874 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890647 4874 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890668 4874 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890690 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890709 4874 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890727 4874 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890749 4874 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890774 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890792 4874 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890811 4874 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890832 4874 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890849 4874 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890868 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890887 4874 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890906 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890927 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890946 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890967 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.890986 4874 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891002 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891022 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891041 4874 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891057 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891074 4874 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891091 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891109 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891127 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891143 4874 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891176 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891194 4874 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891213 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891232 4874 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891251 4874 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891802 4874 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891826 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891847 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891865 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891883 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891901 4874 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891919 4874 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891936 4874 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891953 4874 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891972 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891990 4874 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892009 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892026 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892042 4874 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892062 4874 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892081 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892098 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892116 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892134 4874 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892151 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892167 4874 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892183 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892200 4874 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892217 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892234 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892250 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892669 4874 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892691 4874 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892709 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892726 4874 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892743 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892768 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892787 4874 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892819 4874 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892843 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892863 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892882 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892899 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892918 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892936 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892955 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892972 4874 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.892990 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893008 4874 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893025 4874 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893043 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893062 4874 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893084 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893102 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893122 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893140 4874 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893158 4874 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893176 4874 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893193 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893210 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893228 4874 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893246 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893263 4874 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893281 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893300 4874 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893318 4874 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893338 4874 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893356 4874 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893386 4874 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893443 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893461 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893480 4874 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893498 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893518 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893535 4874 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893553 4874 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893570 4874 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893587 4874 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893604 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893623 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893642 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893661 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893679 4874 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893698 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893717 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893735 4874 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893752 4874 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893769 4874 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893786 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893804 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893820 4874 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893840 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893856 4874 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893872 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891411 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.893918 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.894316 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.898370 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.899602 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.899897 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.891605 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.900569 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.901055 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.901441 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.901604 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.901623 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.901642 4874 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.901736 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:45.401697933 +0000 UTC m=+19.246769213 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.903452 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.904634 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.905128 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.906084 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.906634 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.907172 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.907885 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.907933 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.907955 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.908332 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.908593 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.908723 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.908914 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.908939 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.909310 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.909629 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-4prkg"] Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.909663 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.909958 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.909986 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-pdb2m"] Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.909995 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.910057 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.910223 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.910363 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.910388 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.910421 4874 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.910463 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:44 crc kubenswrapper[4874]: E0122 11:40:44.910504 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:45.410472485 +0000 UTC m=+19.255543755 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.910539 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-krrtc"] Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.910585 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.911133 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.911535 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.911937 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.912811 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.913159 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.913467 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.913464 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.880345 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.913633 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.913633 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.913678 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.913788 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.913910 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.913920 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.913935 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.914049 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.920474 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.920680 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.922617 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.920728 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.921549 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.921748 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.933736 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.933974 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.934696 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.935676 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.936249 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.937776 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.938074 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.938328 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.938570 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.938612 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.939736 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.939769 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.942892 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.942970 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.943147 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.944693 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.944875 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.950168 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.951879 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.956565 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.962073 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.964879 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.978044 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.980455 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.988682 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:44 crc kubenswrapper[4874]: W0122 11:40:44.993890 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-37ae074ff359889d0d2a9d1bb3b631b2a01165b5eedee47c39d5f8871ab33af5 WatchSource:0}: Error finding container 37ae074ff359889d0d2a9d1bb3b631b2a01165b5eedee47c39d5f8871ab33af5: Status 404 returned error can't find the container with id 37ae074ff359889d0d2a9d1bb3b631b2a01165b5eedee47c39d5f8871ab33af5 Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.994997 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7c9653f9-cd5b-4b7a-8056-80ae8235d039-rootfs\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995028 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-cni-binary-copy\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995052 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995071 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-cnibin\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995086 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-conf-dir\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995101 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995117 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-var-lib-kubelet\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995132 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-socket-dir-parent\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995146 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-system-cni-dir\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995161 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-cni-dir\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995175 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-system-cni-dir\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995191 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-run-k8s-cni-cncf-io\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995205 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-run-multus-certs\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995221 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995236 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssgjv\" (UniqueName: \"kubernetes.io/projected/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-kube-api-access-ssgjv\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995255 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-var-lib-cni-multus\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995280 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-run-netns\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995294 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-os-release\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995308 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7c9653f9-cd5b-4b7a-8056-80ae8235d039-mcd-auth-proxy-config\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995324 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-cnibin\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995338 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7c9653f9-cd5b-4b7a-8056-80ae8235d039-proxy-tls\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995357 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f5ca785e-1db4-4e08-9ad0-66158728b48a-hosts-file\") pod \"node-resolver-prbck\" (UID: \"f5ca785e-1db4-4e08-9ad0-66158728b48a\") " pod="openshift-dns/node-resolver-prbck" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995374 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7628z\" (UniqueName: \"kubernetes.io/projected/f5ca785e-1db4-4e08-9ad0-66158728b48a-kube-api-access-7628z\") pod \"node-resolver-prbck\" (UID: \"f5ca785e-1db4-4e08-9ad0-66158728b48a\") " pod="openshift-dns/node-resolver-prbck" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995410 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-var-lib-cni-bin\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995427 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-daemon-config\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995442 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995459 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txncl\" (UniqueName: \"kubernetes.io/projected/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-kube-api-access-txncl\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995482 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-os-release\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995556 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995600 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-cni-binary-copy\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995620 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-etc-kubernetes\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995654 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995667 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmx5h\" (UniqueName: \"kubernetes.io/projected/7c9653f9-cd5b-4b7a-8056-80ae8235d039-kube-api-access-kmx5h\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995684 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f5ca785e-1db4-4e08-9ad0-66158728b48a-hosts-file\") pod \"node-resolver-prbck\" (UID: \"f5ca785e-1db4-4e08-9ad0-66158728b48a\") " pod="openshift-dns/node-resolver-prbck" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.995710 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-hostroot\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996023 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996050 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996064 4874 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996076 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996090 4874 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996105 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996115 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996126 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996135 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996145 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996756 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996774 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996785 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996794 4874 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996804 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996815 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996824 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996833 4874 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996843 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996854 4874 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996864 4874 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996873 4874 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996881 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996889 4874 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996897 4874 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996913 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996922 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996930 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996939 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996948 4874 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996956 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996965 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.996973 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997287 4874 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997298 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997306 4874 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997315 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997324 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997336 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997345 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997354 4874 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997362 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997372 4874 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997382 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997438 4874 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997450 4874 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997459 4874 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997469 4874 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997477 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997486 4874 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997495 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997503 4874 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:44 crc kubenswrapper[4874]: I0122 11:40:44.997513 4874 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:44.997522 4874 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:44.997530 4874 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:44.997539 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:44.997547 4874 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:44.997555 4874 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:44.997564 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:44.997572 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:44.997580 4874 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:44.997588 4874 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:44.997596 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:44.998181 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.009042 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.012804 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7628z\" (UniqueName: \"kubernetes.io/projected/f5ca785e-1db4-4e08-9ad0-66158728b48a-kube-api-access-7628z\") pod \"node-resolver-prbck\" (UID: \"f5ca785e-1db4-4e08-9ad0-66158728b48a\") " pod="openshift-dns/node-resolver-prbck" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.020464 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.036279 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.052674 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.071778 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.082356 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.091250 4874 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.091315 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.091501 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098244 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txncl\" (UniqueName: \"kubernetes.io/projected/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-kube-api-access-txncl\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098297 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-var-lib-cni-bin\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098322 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-daemon-config\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098346 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098374 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-os-release\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098413 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-cni-binary-copy\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098438 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-etc-kubernetes\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098471 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmx5h\" (UniqueName: \"kubernetes.io/projected/7c9653f9-cd5b-4b7a-8056-80ae8235d039-kube-api-access-kmx5h\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098492 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-hostroot\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098514 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7c9653f9-cd5b-4b7a-8056-80ae8235d039-rootfs\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098535 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-cni-binary-copy\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098558 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098582 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-cnibin\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098604 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-conf-dir\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098627 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-var-lib-kubelet\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098651 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-socket-dir-parent\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098674 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-system-cni-dir\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098693 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-cni-dir\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098714 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-system-cni-dir\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098735 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-run-k8s-cni-cncf-io\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098757 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-run-multus-certs\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098779 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssgjv\" (UniqueName: \"kubernetes.io/projected/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-kube-api-access-ssgjv\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098813 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-run-netns\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098835 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-var-lib-cni-multus\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098858 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7c9653f9-cd5b-4b7a-8056-80ae8235d039-mcd-auth-proxy-config\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098882 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-cnibin\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098901 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-os-release\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.098924 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7c9653f9-cd5b-4b7a-8056-80ae8235d039-proxy-tls\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.099319 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-cni-binary-copy\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.099619 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-var-lib-cni-bin\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.099985 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-system-cni-dir\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.099987 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-os-release\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100150 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-conf-dir\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100189 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-var-lib-kubelet\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100241 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-cnibin\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100283 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-run-netns\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100616 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-cni-dir\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100683 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-hostroot\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100625 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-socket-dir-parent\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100684 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-var-lib-cni-multus\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100617 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100728 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-system-cni-dir\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100756 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-run-k8s-cni-cncf-io\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100769 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7c9653f9-cd5b-4b7a-8056-80ae8235d039-rootfs\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100776 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-cnibin\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100795 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-os-release\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100846 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-host-run-multus-certs\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.100864 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-etc-kubernetes\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.101273 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.101387 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-cni-binary-copy\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.101809 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-multus-daemon-config\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.102195 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7c9653f9-cd5b-4b7a-8056-80ae8235d039-mcd-auth-proxy-config\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.103504 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7c9653f9-cd5b-4b7a-8056-80ae8235d039-proxy-tls\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.116927 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmx5h\" (UniqueName: \"kubernetes.io/projected/7c9653f9-cd5b-4b7a-8056-80ae8235d039-kube-api-access-kmx5h\") pod \"machine-config-daemon-4prkg\" (UID: \"7c9653f9-cd5b-4b7a-8056-80ae8235d039\") " pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.117902 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssgjv\" (UniqueName: \"kubernetes.io/projected/977746b5-ac1b-4b6e-bdbc-ddd90225e68c-kube-api-access-ssgjv\") pod \"multus-krrtc\" (UID: \"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\") " pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.119936 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txncl\" (UniqueName: \"kubernetes.io/projected/0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4-kube-api-access-txncl\") pod \"multus-additional-cni-plugins-pdb2m\" (UID: \"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\") " pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.243315 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.251918 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-krrtc" Jan 22 11:40:45 crc kubenswrapper[4874]: W0122 11:40:45.257431 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c9653f9_cd5b_4b7a_8056_80ae8235d039.slice/crio-3ba93b22c47310b71d0d933debf3f1ccffae2c2dd7f4d7bbbbad4f798ce7c1a0 WatchSource:0}: Error finding container 3ba93b22c47310b71d0d933debf3f1ccffae2c2dd7f4d7bbbbad4f798ce7c1a0: Status 404 returned error can't find the container with id 3ba93b22c47310b71d0d933debf3f1ccffae2c2dd7f4d7bbbbad4f798ce7c1a0 Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.262125 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.265116 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6tmll"] Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.266226 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: W0122 11:40:45.266713 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod977746b5_ac1b_4b6e_bdbc_ddd90225e68c.slice/crio-15e7294d175437ddde2823bdef9219465ecc4f5bc18338a73afd1404b3637ca7 WatchSource:0}: Error finding container 15e7294d175437ddde2823bdef9219465ecc4f5bc18338a73afd1404b3637ca7: Status 404 returned error can't find the container with id 15e7294d175437ddde2823bdef9219465ecc4f5bc18338a73afd1404b3637ca7 Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.270901 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.270910 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.270968 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.270924 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.271061 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.271105 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.271374 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.273651 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 11:40:45 crc kubenswrapper[4874]: W0122 11:40:45.280008 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dbaa8a2_585e_4ab7_a8f5_7b224ad8c5c4.slice/crio-cfd85d40370beef355d62618552b93b7c8b9c6377a7e2284aad1ac33a17abc09 WatchSource:0}: Error finding container cfd85d40370beef355d62618552b93b7c8b9c6377a7e2284aad1ac33a17abc09: Status 404 returned error can't find the container with id cfd85d40370beef355d62618552b93b7c8b9c6377a7e2284aad1ac33a17abc09 Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.288035 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.292332 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-prbck" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.299974 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-slash\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300010 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-openvswitch\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300028 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovn-node-metrics-cert\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300047 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-netns\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300140 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-netd\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300183 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-etc-openvswitch\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300221 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-ovn\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300244 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-script-lib\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300272 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-kubelet\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300286 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-systemd-units\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300301 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-systemd\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300323 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-node-log\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300345 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-config\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300370 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9lv4\" (UniqueName: \"kubernetes.io/projected/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-kube-api-access-t9lv4\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300443 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-bin\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300527 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-log-socket\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300551 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300592 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-var-lib-openvswitch\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300628 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-ovn-kubernetes\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.300650 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-env-overrides\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.306045 4874 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-22 11:35:44 +0000 UTC, rotation deadline is 2026-10-10 20:37:33.202021591 +0000 UTC Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.306117 4874 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6272h56m47.895908153s for next certificate rotation Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.307242 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: W0122 11:40:45.316165 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-6083a98e963c10d34216819fff1fe8e431382ea80b86b0b07e7683f96ecfcf58 WatchSource:0}: Error finding container 6083a98e963c10d34216819fff1fe8e431382ea80b86b0b07e7683f96ecfcf58: Status 404 returned error can't find the container with id 6083a98e963c10d34216819fff1fe8e431382ea80b86b0b07e7683f96ecfcf58 Jan 22 11:40:45 crc kubenswrapper[4874]: W0122 11:40:45.319952 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-a4229c969216bf4a1eeb7c04ec3d2e7ee8178a7f38ade5781ca125e141ed5cff WatchSource:0}: Error finding container a4229c969216bf4a1eeb7c04ec3d2e7ee8178a7f38ade5781ca125e141ed5cff: Status 404 returned error can't find the container with id a4229c969216bf4a1eeb7c04ec3d2e7ee8178a7f38ade5781ca125e141ed5cff Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.323275 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.347843 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.365506 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.380808 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.394904 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.401633 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.401740 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.401772 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-log-socket\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.401795 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.401915 4874 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402013 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.401913 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-log-socket\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.401916 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:40:46.401896843 +0000 UTC m=+20.246967913 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402072 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402099 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-var-lib-openvswitch\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402123 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402144 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-ovn-kubernetes\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402162 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-env-overrides\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402182 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-slash\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402200 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-openvswitch\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402221 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovn-node-metrics-cert\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402243 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-netns\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402265 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-netd\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402283 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-etc-openvswitch\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.402303 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.402322 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402327 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-ovn\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.402338 4874 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402360 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-ovn-kubernetes\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.402372 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:46.402360418 +0000 UTC m=+20.247431488 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.402431 4874 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.402462 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:46.40245188 +0000 UTC m=+20.247522950 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.402475 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:46.402468961 +0000 UTC m=+20.247540031 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402501 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-var-lib-openvswitch\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402303 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-ovn\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402562 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-script-lib\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402596 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-kubelet\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402614 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-systemd-units\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402637 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-systemd\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402657 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-node-log\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402684 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-config\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402703 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9lv4\" (UniqueName: \"kubernetes.io/projected/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-kube-api-access-t9lv4\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402733 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-bin\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402819 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-bin\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402916 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-env-overrides\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402961 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-netns\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402964 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-systemd\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.402998 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-netd\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.403001 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-etc-openvswitch\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.403032 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-kubelet\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.403038 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-systemd-units\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.403085 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-node-log\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.403119 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-slash\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.403494 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-openvswitch\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.403609 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-script-lib\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.408688 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-config\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.413388 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.417780 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovn-node-metrics-cert\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.424119 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9lv4\" (UniqueName: \"kubernetes.io/projected/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-kube-api-access-t9lv4\") pod \"ovnkube-node-6tmll\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.430615 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.443003 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.458945 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.473780 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.491309 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.504238 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.505523 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.505558 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.505571 4874 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:45 crc kubenswrapper[4874]: E0122 11:40:45.505624 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:46.505609753 +0000 UTC m=+20.350680823 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.578383 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.669845 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 00:08:42.838200569 +0000 UTC Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.818054 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-prbck" event={"ID":"f5ca785e-1db4-4e08-9ad0-66158728b48a","Type":"ContainerStarted","Data":"aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.818348 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-prbck" event={"ID":"f5ca785e-1db4-4e08-9ad0-66158728b48a","Type":"ContainerStarted","Data":"5ee10dc6fb3000354d81b9f4fbbcb513ac5ba37ac6db6b8ed3f180cb62d9e8ea"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.818995 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-krrtc" event={"ID":"977746b5-ac1b-4b6e-bdbc-ddd90225e68c","Type":"ContainerStarted","Data":"600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.819021 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-krrtc" event={"ID":"977746b5-ac1b-4b6e-bdbc-ddd90225e68c","Type":"ContainerStarted","Data":"15e7294d175437ddde2823bdef9219465ecc4f5bc18338a73afd1404b3637ca7"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.819752 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a4229c969216bf4a1eeb7c04ec3d2e7ee8178a7f38ade5781ca125e141ed5cff"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.821273 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.822621 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.822867 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.823623 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.823697 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6083a98e963c10d34216819fff1fe8e431382ea80b86b0b07e7683f96ecfcf58"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.824826 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.824854 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.824869 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"37ae074ff359889d0d2a9d1bb3b631b2a01165b5eedee47c39d5f8871ab33af5"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.826274 4874 generic.go:334] "Generic (PLEG): container finished" podID="0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4" containerID="b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a" exitCode=0 Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.826337 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" event={"ID":"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4","Type":"ContainerDied","Data":"b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.826356 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" event={"ID":"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4","Type":"ContainerStarted","Data":"cfd85d40370beef355d62618552b93b7c8b9c6377a7e2284aad1ac33a17abc09"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.827889 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.827918 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.827932 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"3ba93b22c47310b71d0d933debf3f1ccffae2c2dd7f4d7bbbbad4f798ce7c1a0"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.831903 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f" exitCode=0 Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.831953 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.831980 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"1a325212afdff76674e62ed80b3cf828c221bb977e244657ca644f4020d804af"} Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.835858 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.846726 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.859077 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.872154 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.883869 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.896843 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.910741 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.925045 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.940371 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.960753 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.975957 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.987201 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:45 crc kubenswrapper[4874]: I0122 11:40:45.997756 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:45Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.015248 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.026026 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.038686 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.052036 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.068320 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.081486 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.147872 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.180222 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.200427 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.214562 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.229946 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.350111 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.362556 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.364083 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.378164 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.395051 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.413060 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.413169 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.413254 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:40:48.413225173 +0000 UTC m=+22.258296243 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.413268 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.413284 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.413294 4874 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.413368 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:48.413353347 +0000 UTC m=+22.258424417 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.413443 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.413469 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.413512 4874 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.413533 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:48.413526302 +0000 UTC m=+22.258597372 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.413614 4874 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.413681 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:48.413669197 +0000 UTC m=+22.258740317 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.415989 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.428474 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.441155 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.455631 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.472338 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.490526 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.506217 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.514041 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.514254 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.514448 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.514464 4874 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.514522 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:48.514504374 +0000 UTC m=+22.359575514 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.523690 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.547717 4874 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.548973 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-operator/pods/iptables-alerter-4ln5h/status\": read tcp 38.102.83.153:52796->38.102.83.153:6443: use of closed network connection" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.588490 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.636827 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.669192 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.670157 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 10:24:22.283485033 +0000 UTC Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.710106 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.715238 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.715242 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.715429 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.715252 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.715683 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:40:46 crc kubenswrapper[4874]: E0122 11:40:46.715497 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.719932 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.720801 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.721670 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.722367 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.723061 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.724716 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.725504 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.726589 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.727437 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.727980 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.728905 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.729642 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.730635 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.731169 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.732390 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.732895 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.734325 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.734783 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.735309 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.736899 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.737444 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.737972 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.740088 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.740876 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.741756 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.742347 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.743532 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.743990 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.745015 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.745550 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.745990 4874 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.746082 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.748053 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.748791 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.749363 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.749921 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.751347 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.752022 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.753041 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.753826 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.754836 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.755308 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.756240 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.756910 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.757937 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.758367 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.759370 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.759914 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.761120 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.761747 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.762841 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.763364 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.763990 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.765174 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.765757 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.793671 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.829594 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.836590 4874 generic.go:334] "Generic (PLEG): container finished" podID="0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4" containerID="d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a" exitCode=0 Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.836662 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" event={"ID":"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4","Type":"ContainerDied","Data":"d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a"} Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.840307 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2"} Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.840359 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348"} Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.840370 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba"} Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.840383 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794"} Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.874167 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.908831 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.948627 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:46 crc kubenswrapper[4874]: I0122 11:40:46.990432 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.030850 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.068646 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.120525 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.151513 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.192082 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.241144 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.273152 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.331427 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.365425 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.390568 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.429749 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.451725 4874 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.453354 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.453441 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.453455 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.453568 4874 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.467122 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.521809 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-q2rnk"] Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.522032 4874 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.522206 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-q2rnk" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.522312 4874 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.523285 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.523316 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.523323 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.523335 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.523344 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:47Z","lastTransitionTime":"2026-01-22T11:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:47 crc kubenswrapper[4874]: E0122 11:40:47.539688 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.543285 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.543329 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.543339 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.543358 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.543368 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:47Z","lastTransitionTime":"2026-01-22T11:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.547654 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: E0122 11:40:47.554608 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.560783 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.560834 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.560850 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.560869 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.560834 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.560881 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:47Z","lastTransitionTime":"2026-01-22T11:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:47 crc kubenswrapper[4874]: E0122 11:40:47.573091 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.576435 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.576463 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.576472 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.576487 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.576496 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:47Z","lastTransitionTime":"2026-01-22T11:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.580844 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 22 11:40:47 crc kubenswrapper[4874]: E0122 11:40:47.589198 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.592234 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.592278 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.592290 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.592309 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.592321 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:47Z","lastTransitionTime":"2026-01-22T11:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.600613 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 22 11:40:47 crc kubenswrapper[4874]: E0122 11:40:47.606561 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: E0122 11:40:47.606675 4874 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.608884 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.608930 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.608945 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.608969 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.608982 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:47Z","lastTransitionTime":"2026-01-22T11:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.621251 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.623789 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/561c457d-4767-4b66-a07a-c435b7c9f161-host\") pod \"node-ca-q2rnk\" (UID: \"561c457d-4767-4b66-a07a-c435b7c9f161\") " pod="openshift-image-registry/node-ca-q2rnk" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.623817 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/561c457d-4767-4b66-a07a-c435b7c9f161-serviceca\") pod \"node-ca-q2rnk\" (UID: \"561c457d-4767-4b66-a07a-c435b7c9f161\") " pod="openshift-image-registry/node-ca-q2rnk" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.623836 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn6zq\" (UniqueName: \"kubernetes.io/projected/561c457d-4767-4b66-a07a-c435b7c9f161-kube-api-access-dn6zq\") pod \"node-ca-q2rnk\" (UID: \"561c457d-4767-4b66-a07a-c435b7c9f161\") " pod="openshift-image-registry/node-ca-q2rnk" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.671150 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 15:22:24.811162877 +0000 UTC Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.672712 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.711823 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.711870 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.711883 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.711901 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.711912 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:47Z","lastTransitionTime":"2026-01-22T11:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.716689 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.724530 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn6zq\" (UniqueName: \"kubernetes.io/projected/561c457d-4767-4b66-a07a-c435b7c9f161-kube-api-access-dn6zq\") pod \"node-ca-q2rnk\" (UID: \"561c457d-4767-4b66-a07a-c435b7c9f161\") " pod="openshift-image-registry/node-ca-q2rnk" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.724581 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/561c457d-4767-4b66-a07a-c435b7c9f161-host\") pod \"node-ca-q2rnk\" (UID: \"561c457d-4767-4b66-a07a-c435b7c9f161\") " pod="openshift-image-registry/node-ca-q2rnk" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.724604 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/561c457d-4767-4b66-a07a-c435b7c9f161-serviceca\") pod \"node-ca-q2rnk\" (UID: \"561c457d-4767-4b66-a07a-c435b7c9f161\") " pod="openshift-image-registry/node-ca-q2rnk" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.724754 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/561c457d-4767-4b66-a07a-c435b7c9f161-host\") pod \"node-ca-q2rnk\" (UID: \"561c457d-4767-4b66-a07a-c435b7c9f161\") " pod="openshift-image-registry/node-ca-q2rnk" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.725723 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/561c457d-4767-4b66-a07a-c435b7c9f161-serviceca\") pod \"node-ca-q2rnk\" (UID: \"561c457d-4767-4b66-a07a-c435b7c9f161\") " pod="openshift-image-registry/node-ca-q2rnk" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.753754 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.788924 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn6zq\" (UniqueName: \"kubernetes.io/projected/561c457d-4767-4b66-a07a-c435b7c9f161-kube-api-access-dn6zq\") pod \"node-ca-q2rnk\" (UID: \"561c457d-4767-4b66-a07a-c435b7c9f161\") " pod="openshift-image-registry/node-ca-q2rnk" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.810890 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.815599 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.815641 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.815651 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.815670 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.815681 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:47Z","lastTransitionTime":"2026-01-22T11:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.833542 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-q2rnk" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.845961 4874 generic.go:334] "Generic (PLEG): container finished" podID="0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4" containerID="a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14" exitCode=0 Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.846050 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" event={"ID":"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4","Type":"ContainerDied","Data":"a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14"} Jan 22 11:40:47 crc kubenswrapper[4874]: W0122 11:40:47.849240 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod561c457d_4767_4b66_a07a_c435b7c9f161.slice/crio-da7d48bcb9a2d643d27a44577a44c56b3bb28ef9b367b7558dc1ebdf9a12f7c8 WatchSource:0}: Error finding container da7d48bcb9a2d643d27a44577a44c56b3bb28ef9b367b7558dc1ebdf9a12f7c8: Status 404 returned error can't find the container with id da7d48bcb9a2d643d27a44577a44c56b3bb28ef9b367b7558dc1ebdf9a12f7c8 Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.851527 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.855127 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079"} Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.855172 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6"} Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.898752 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.919626 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.919688 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.919699 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.919717 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.919729 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:47Z","lastTransitionTime":"2026-01-22T11:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.932173 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:47 crc kubenswrapper[4874]: I0122 11:40:47.971252 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.010611 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.021816 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.021849 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.021860 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.021872 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.021881 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:48Z","lastTransitionTime":"2026-01-22T11:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.049661 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.090518 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.124193 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.124235 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.124245 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.124258 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.124269 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:48Z","lastTransitionTime":"2026-01-22T11:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.130697 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.171038 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.217261 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.227049 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.227083 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.227095 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.227112 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.227125 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:48Z","lastTransitionTime":"2026-01-22T11:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.258087 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.300936 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.330027 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.330071 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.330084 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.330131 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.330148 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:48Z","lastTransitionTime":"2026-01-22T11:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.330337 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.373504 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.416163 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.430878 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.430987 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:40:52.430969746 +0000 UTC m=+26.276040826 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.431044 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.431072 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.431094 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.431190 4874 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.431201 4874 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.431218 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.431238 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.431249 4874 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.431251 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:52.431233844 +0000 UTC m=+26.276304934 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.431285 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:52.431262745 +0000 UTC m=+26.276333825 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.431307 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:52.431297497 +0000 UTC m=+26.276368577 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.432871 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.432907 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.432919 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.432936 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.432947 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:48Z","lastTransitionTime":"2026-01-22T11:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.449105 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.492534 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.532131 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.532330 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.532352 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.532374 4874 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.532519 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:52.532496485 +0000 UTC m=+26.377567565 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.535479 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.537353 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.537471 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.537502 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.537533 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.537555 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:48Z","lastTransitionTime":"2026-01-22T11:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.583075 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.611246 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.640638 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.640672 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.640681 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.640695 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.640706 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:48Z","lastTransitionTime":"2026-01-22T11:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.651530 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.671645 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 08:52:34.192009612 +0000 UTC Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.690988 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.715347 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.715550 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.715632 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.715672 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.715801 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:40:48 crc kubenswrapper[4874]: E0122 11:40:48.715942 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.741638 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.743214 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.743244 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.743258 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.743274 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.743286 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:48Z","lastTransitionTime":"2026-01-22T11:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.767892 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.810311 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.846181 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.846269 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.846293 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.846325 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.846349 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:48Z","lastTransitionTime":"2026-01-22T11:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.859344 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.862731 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.865672 4874 generic.go:334] "Generic (PLEG): container finished" podID="0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4" containerID="6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63" exitCode=0 Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.865730 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" event={"ID":"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4","Type":"ContainerDied","Data":"6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.866845 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-q2rnk" event={"ID":"561c457d-4767-4b66-a07a-c435b7c9f161","Type":"ContainerStarted","Data":"a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.866869 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-q2rnk" event={"ID":"561c457d-4767-4b66-a07a-c435b7c9f161","Type":"ContainerStarted","Data":"da7d48bcb9a2d643d27a44577a44c56b3bb28ef9b367b7558dc1ebdf9a12f7c8"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.887907 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.929812 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.948385 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.948435 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.948447 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.948460 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.948469 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:48Z","lastTransitionTime":"2026-01-22T11:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:48 crc kubenswrapper[4874]: I0122 11:40:48.969942 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:48Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.015731 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.049831 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.051202 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.051268 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.051284 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.051304 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.051318 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:49Z","lastTransitionTime":"2026-01-22T11:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.088282 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.128441 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.153635 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.153674 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.153685 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.153701 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.153712 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:49Z","lastTransitionTime":"2026-01-22T11:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.168946 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.212353 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.249802 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.255273 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.255304 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.255313 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.255325 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.255334 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:49Z","lastTransitionTime":"2026-01-22T11:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.290024 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.337930 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.358141 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.358176 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.358187 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.358209 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.358236 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:49Z","lastTransitionTime":"2026-01-22T11:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.376370 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.409244 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.453553 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.460445 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.460495 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.460509 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.460527 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.460539 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:49Z","lastTransitionTime":"2026-01-22T11:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.563360 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.563414 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.563429 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.563446 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.563457 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:49Z","lastTransitionTime":"2026-01-22T11:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.666335 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.666429 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.666449 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.666476 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.666493 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:49Z","lastTransitionTime":"2026-01-22T11:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.671871 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 10:28:03.143420805 +0000 UTC Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.768269 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.768308 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.768318 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.768334 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.768347 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:49Z","lastTransitionTime":"2026-01-22T11:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.871037 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.871067 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.871075 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.871087 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.871097 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:49Z","lastTransitionTime":"2026-01-22T11:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.873213 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.875918 4874 generic.go:334] "Generic (PLEG): container finished" podID="0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4" containerID="cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7" exitCode=0 Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.876272 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" event={"ID":"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4","Type":"ContainerDied","Data":"cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.891279 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.906086 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.928165 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.943740 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.955147 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.965899 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.976144 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.976189 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.976201 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.976218 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.976232 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:49Z","lastTransitionTime":"2026-01-22T11:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.980630 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:49 crc kubenswrapper[4874]: I0122 11:40:49.995939 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.005956 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.014919 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.027537 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.040437 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.057578 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.068518 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.078219 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.078329 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.078445 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.078534 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.078624 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:50Z","lastTransitionTime":"2026-01-22T11:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.180491 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.180532 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.180542 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.180557 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.180568 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:50Z","lastTransitionTime":"2026-01-22T11:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.283660 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.283697 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.283709 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.283726 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.283738 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:50Z","lastTransitionTime":"2026-01-22T11:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.386281 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.386317 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.386326 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.386341 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.386350 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:50Z","lastTransitionTime":"2026-01-22T11:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.489177 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.489206 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.489216 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.489229 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.489239 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:50Z","lastTransitionTime":"2026-01-22T11:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.592653 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.592705 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.592717 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.592737 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.592748 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:50Z","lastTransitionTime":"2026-01-22T11:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.673008 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 07:13:06.021989114 +0000 UTC Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.695047 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.695093 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.695105 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.695124 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.695137 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:50Z","lastTransitionTime":"2026-01-22T11:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.715541 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.715559 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:50 crc kubenswrapper[4874]: E0122 11:40:50.715758 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:40:50 crc kubenswrapper[4874]: E0122 11:40:50.715855 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.715573 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:50 crc kubenswrapper[4874]: E0122 11:40:50.715961 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.797649 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.797714 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.797729 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.797753 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.797769 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:50Z","lastTransitionTime":"2026-01-22T11:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.882642 4874 generic.go:334] "Generic (PLEG): container finished" podID="0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4" containerID="617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94" exitCode=0 Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.882681 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" event={"ID":"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4","Type":"ContainerDied","Data":"617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94"} Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.900226 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.900261 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.900269 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.900284 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.900293 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:50Z","lastTransitionTime":"2026-01-22T11:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.902026 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.917270 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.937747 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.951963 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.969184 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:50 crc kubenswrapper[4874]: I0122 11:40:50.988083 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.000643 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:50Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.002477 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.002511 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.002521 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.002544 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.002561 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:51Z","lastTransitionTime":"2026-01-22T11:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.012828 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.026325 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.040218 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.051171 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.060989 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.073127 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.092986 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.104744 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.104792 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.104803 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.104821 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.104831 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:51Z","lastTransitionTime":"2026-01-22T11:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.211030 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.211100 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.211184 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.211223 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.211237 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:51Z","lastTransitionTime":"2026-01-22T11:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.315377 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.315730 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.315741 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.315757 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.315766 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:51Z","lastTransitionTime":"2026-01-22T11:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.417851 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.417889 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.417900 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.417916 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.417926 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:51Z","lastTransitionTime":"2026-01-22T11:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.520256 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.520280 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.520288 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.520300 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.520309 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:51Z","lastTransitionTime":"2026-01-22T11:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.622742 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.622792 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.622804 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.622822 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.622834 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:51Z","lastTransitionTime":"2026-01-22T11:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.673783 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 04:51:13.144057122 +0000 UTC Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.725201 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.725244 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.725253 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.725270 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.725280 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:51Z","lastTransitionTime":"2026-01-22T11:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.827829 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.827879 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.827891 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.827908 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.827920 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:51Z","lastTransitionTime":"2026-01-22T11:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.889172 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.889413 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.889477 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.894814 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" event={"ID":"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4","Type":"ContainerStarted","Data":"b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.904565 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.911647 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.913349 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.917554 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.929823 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.929852 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.929862 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.929876 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.929887 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:51Z","lastTransitionTime":"2026-01-22T11:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.933982 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.944044 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.954302 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.966574 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.975546 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:51 crc kubenswrapper[4874]: I0122 11:40:51.990737 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.001304 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:51Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.010309 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.020913 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.031892 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.031932 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.031949 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.031974 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.032011 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:52Z","lastTransitionTime":"2026-01-22T11:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.033889 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.051438 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.062768 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.076826 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.094489 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.098422 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.100728 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.103194 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.113532 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.126455 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.135246 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.135285 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.135293 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.135310 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.135320 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:52Z","lastTransitionTime":"2026-01-22T11:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.140212 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.158512 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.170631 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.181767 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.193695 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.201970 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.212082 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.221338 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.229734 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.238310 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.238345 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.238354 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.238371 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.238381 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:52Z","lastTransitionTime":"2026-01-22T11:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.239627 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.250801 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.261283 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.273586 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.282501 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.292462 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.303649 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.313722 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.323252 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.338013 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.340668 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.340700 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.340710 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.340725 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.340735 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:52Z","lastTransitionTime":"2026-01-22T11:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.358820 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.369585 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.381132 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.392759 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.402667 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.417985 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:52Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.442756 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.442794 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.442823 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.442838 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.442850 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:52Z","lastTransitionTime":"2026-01-22T11:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.471378 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.471537 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.471585 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.471625 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:41:00.471590765 +0000 UTC m=+34.316661845 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.471653 4874 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.471678 4874 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.471690 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.471710 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:00.471694539 +0000 UTC m=+34.316765659 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.471728 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:00.471718959 +0000 UTC m=+34.316790149 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.471960 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.472021 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.472048 4874 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.472165 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:00.472132523 +0000 UTC m=+34.317203643 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.508310 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.545159 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.545195 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.545204 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.545216 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.545224 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:52Z","lastTransitionTime":"2026-01-22T11:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.573077 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.573437 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.573494 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.573521 4874 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.573612 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:00.573589641 +0000 UTC m=+34.418660751 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.648177 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.648226 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.648236 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.648251 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.648262 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:52Z","lastTransitionTime":"2026-01-22T11:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.674097 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 10:10:28.206002189 +0000 UTC Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.715547 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.715626 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.715566 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.715755 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.715684 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.715845 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.750512 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.750557 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.750567 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.750580 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.750590 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:52Z","lastTransitionTime":"2026-01-22T11:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.852975 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.853023 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.853031 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.853046 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.853055 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:52Z","lastTransitionTime":"2026-01-22T11:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:52 crc kubenswrapper[4874]: E0122 11:40:52.904078 4874 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.956324 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.956371 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.956387 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.956437 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:52 crc kubenswrapper[4874]: I0122 11:40:52.956461 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:52Z","lastTransitionTime":"2026-01-22T11:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.059573 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.059623 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.059634 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.059650 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.059662 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:53Z","lastTransitionTime":"2026-01-22T11:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.163037 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.163090 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.163102 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.163120 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.163133 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:53Z","lastTransitionTime":"2026-01-22T11:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.265457 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.265498 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.265512 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.265528 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.265538 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:53Z","lastTransitionTime":"2026-01-22T11:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.368561 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.368604 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.368617 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.368633 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.368646 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:53Z","lastTransitionTime":"2026-01-22T11:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.474841 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.474883 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.474892 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.474913 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.474930 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:53Z","lastTransitionTime":"2026-01-22T11:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.577657 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.577697 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.577707 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.577725 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.577738 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:53Z","lastTransitionTime":"2026-01-22T11:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.674311 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 17:20:46.564312986 +0000 UTC Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.681619 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.681690 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.681706 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.681724 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.681754 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:53Z","lastTransitionTime":"2026-01-22T11:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.785028 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.785082 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.785096 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.785116 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.785131 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:53Z","lastTransitionTime":"2026-01-22T11:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.887660 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.887703 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.887711 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.887725 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.887735 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:53Z","lastTransitionTime":"2026-01-22T11:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.989723 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.989777 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.989789 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.989804 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:53 crc kubenswrapper[4874]: I0122 11:40:53.990126 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:53Z","lastTransitionTime":"2026-01-22T11:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.093457 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.093498 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.093510 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.093527 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.093538 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:54Z","lastTransitionTime":"2026-01-22T11:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.196045 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.196086 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.196095 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.196109 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.196119 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:54Z","lastTransitionTime":"2026-01-22T11:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.299892 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.299946 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.299956 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.299987 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.299999 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:54Z","lastTransitionTime":"2026-01-22T11:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.402712 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.402767 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.402779 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.402800 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.402815 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:54Z","lastTransitionTime":"2026-01-22T11:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.505490 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.505545 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.505557 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.505576 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.505589 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:54Z","lastTransitionTime":"2026-01-22T11:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.608500 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.608551 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.608569 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.608592 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.608608 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:54Z","lastTransitionTime":"2026-01-22T11:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.674547 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 07:32:47.479058154 +0000 UTC Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.712019 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.712088 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.712106 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.712130 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.712147 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:54Z","lastTransitionTime":"2026-01-22T11:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.715177 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.715354 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:54 crc kubenswrapper[4874]: E0122 11:40:54.715369 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.715418 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:54 crc kubenswrapper[4874]: E0122 11:40:54.715489 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:40:54 crc kubenswrapper[4874]: E0122 11:40:54.715572 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.814741 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.814789 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.814802 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.814822 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.814837 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:54Z","lastTransitionTime":"2026-01-22T11:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.905555 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/0.log" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.909050 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8" exitCode=1 Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.909100 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8"} Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.909939 4874 scope.go:117] "RemoveContainer" containerID="72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.917223 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.917252 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.917261 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.917276 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.917286 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:54Z","lastTransitionTime":"2026-01-22T11:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.920566 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:54Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.929807 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:54Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.942105 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:54Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.951184 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:54Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.961254 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:54Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:54 crc kubenswrapper[4874]: I0122 11:40:54.979500 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:54Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.004731 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.018371 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.020053 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.020086 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.020095 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.020110 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.020121 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:55Z","lastTransitionTime":"2026-01-22T11:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.033210 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.043716 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.063523 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:54Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298122 6154 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298314 6154 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 11:40:54.299178 6154 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:40:54.299220 6154 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 11:40:54.299250 6154 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:40:54.299337 6154 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 11:40:54.299719 6154 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:40:54.299899 6154 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 11:40:54.299923 6154 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:54.299953 6154 factory.go:656] Stopping watch factory\\\\nI0122 11:40:54.299985 6154 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:54.300031 6154 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:54.300052 6154 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.075328 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.088950 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.100912 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.115969 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.123026 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.123060 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.123068 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.123085 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.123094 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:55Z","lastTransitionTime":"2026-01-22T11:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.225293 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.225336 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.225351 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.225373 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.225390 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:55Z","lastTransitionTime":"2026-01-22T11:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.327976 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.328027 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.328038 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.328055 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.328066 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:55Z","lastTransitionTime":"2026-01-22T11:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.447719 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.447758 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.447766 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.447783 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.447793 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:55Z","lastTransitionTime":"2026-01-22T11:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.550253 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.550298 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.550310 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.550326 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.550339 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:55Z","lastTransitionTime":"2026-01-22T11:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.652938 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.652971 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.652981 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.652998 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.653010 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:55Z","lastTransitionTime":"2026-01-22T11:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.675579 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 10:52:43.20206447 +0000 UTC Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.755691 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.755742 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.755757 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.755818 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.755836 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:55Z","lastTransitionTime":"2026-01-22T11:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.858184 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.858234 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.858246 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.858265 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.858276 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:55Z","lastTransitionTime":"2026-01-22T11:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.915430 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/0.log" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.919045 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9"} Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.919439 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.931711 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.942283 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.958629 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:54Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298122 6154 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298314 6154 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 11:40:54.299178 6154 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:40:54.299220 6154 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 11:40:54.299250 6154 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:40:54.299337 6154 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 11:40:54.299719 6154 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:40:54.299899 6154 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 11:40:54.299923 6154 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:54.299953 6154 factory.go:656] Stopping watch factory\\\\nI0122 11:40:54.299985 6154 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:54.300031 6154 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:54.300052 6154 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.960725 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.960765 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.960777 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.960795 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.960808 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:55Z","lastTransitionTime":"2026-01-22T11:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.977207 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:55 crc kubenswrapper[4874]: I0122 11:40:55.986733 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.000022 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:55Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.012364 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.022926 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.033865 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.044157 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.054316 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.063240 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.063293 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.063304 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.063319 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.063329 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:56Z","lastTransitionTime":"2026-01-22T11:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.067195 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.085387 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.098337 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.111713 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.165901 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.165940 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.165949 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.165964 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.165974 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:56Z","lastTransitionTime":"2026-01-22T11:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.268886 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.268915 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.268924 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.268938 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.268949 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:56Z","lastTransitionTime":"2026-01-22T11:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.371769 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.371807 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.371816 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.371829 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.371840 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:56Z","lastTransitionTime":"2026-01-22T11:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.474137 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.474181 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.474189 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.474203 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.474212 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:56Z","lastTransitionTime":"2026-01-22T11:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.577672 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.577727 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.577743 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.577766 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.577783 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:56Z","lastTransitionTime":"2026-01-22T11:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.675748 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 19:18:52.744216364 +0000 UTC Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.680647 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.680678 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.680687 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.680702 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.680713 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:56Z","lastTransitionTime":"2026-01-22T11:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.716132 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.716495 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.716489 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:56 crc kubenswrapper[4874]: E0122 11:40:56.716361 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:40:56 crc kubenswrapper[4874]: E0122 11:40:56.717694 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:40:56 crc kubenswrapper[4874]: E0122 11:40:56.717975 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.728919 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.741689 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.752975 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.765918 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.779927 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.782764 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.783060 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.784066 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.784216 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.784347 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:56Z","lastTransitionTime":"2026-01-22T11:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.801002 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.816989 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.830643 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.843358 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.854771 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.871001 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:54Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298122 6154 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298314 6154 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 11:40:54.299178 6154 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:40:54.299220 6154 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 11:40:54.299250 6154 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:40:54.299337 6154 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 11:40:54.299719 6154 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:40:54.299899 6154 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 11:40:54.299923 6154 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:54.299953 6154 factory.go:656] Stopping watch factory\\\\nI0122 11:40:54.299985 6154 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:54.300031 6154 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:54.300052 6154 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.884719 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.886335 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.886431 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.886447 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.886465 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.886500 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:56Z","lastTransitionTime":"2026-01-22T11:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.899009 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.914796 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.922903 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/1.log" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.923510 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/0.log" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.925955 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9" exitCode=1 Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.925995 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9"} Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.926044 4874 scope.go:117] "RemoveContainer" containerID="72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.926643 4874 scope.go:117] "RemoveContainer" containerID="6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.926698 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: E0122 11:40:56.926909 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.938302 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.955712 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.968568 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.980697 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.988929 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.988968 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.988986 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.989002 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.989014 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:56Z","lastTransitionTime":"2026-01-22T11:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:56 crc kubenswrapper[4874]: I0122 11:40:56.992819 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.003334 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.022518 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:54Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298122 6154 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298314 6154 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 11:40:54.299178 6154 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:40:54.299220 6154 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 11:40:54.299250 6154 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:40:54.299337 6154 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 11:40:54.299719 6154 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:40:54.299899 6154 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 11:40:54.299923 6154 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:54.299953 6154 factory.go:656] Stopping watch factory\\\\nI0122 11:40:54.299985 6154 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:54.300031 6154 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:54.300052 6154 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:56Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI0122 11:40:55.854329 6276 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:55.854364 6276 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 11:40:55.854411 6276 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 11:40:55.854421 6276 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:55.854565 6276 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854637 6276 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0122 11:40:55.854653 6276 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:55.854670 6276 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 11:40:55.854674 6276 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 11:40:55.854687 6276 factory.go:656] Stopping watch factory\\\\nI0122 11:40:55.854697 6276 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:55.854712 6276 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854735 6276 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 11:40:55.854740 6276 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0122 11:40:55.854749 6276 handler.go:208] Removed *v1.Pod event handler 6\\\\nF0122 11:40:55.854804 6276 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.033439 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.063025 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.075775 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.090888 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5"] Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.091319 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.092121 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.092158 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.092168 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.092186 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.092198 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.093793 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.093870 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.095132 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.105005 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.115156 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.125066 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.136477 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.145634 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.154211 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.164129 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.175095 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.191738 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.194630 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.194732 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.194746 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.194768 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.194790 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.207069 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.221969 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8470cb5-cfaf-4760-8c07-ce375052950f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.222041 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8470cb5-cfaf-4760-8c07-ce375052950f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.222067 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8470cb5-cfaf-4760-8c07-ce375052950f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.222252 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9smw\" (UniqueName: \"kubernetes.io/projected/a8470cb5-cfaf-4760-8c07-ce375052950f-kube-api-access-t9smw\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.222705 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.235657 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.250209 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.269930 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:54Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298122 6154 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298314 6154 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 11:40:54.299178 6154 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:40:54.299220 6154 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 11:40:54.299250 6154 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:40:54.299337 6154 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 11:40:54.299719 6154 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:40:54.299899 6154 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 11:40:54.299923 6154 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:54.299953 6154 factory.go:656] Stopping watch factory\\\\nI0122 11:40:54.299985 6154 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:54.300031 6154 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:54.300052 6154 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:56Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI0122 11:40:55.854329 6276 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:55.854364 6276 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 11:40:55.854411 6276 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 11:40:55.854421 6276 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:55.854565 6276 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854637 6276 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0122 11:40:55.854653 6276 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:55.854670 6276 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 11:40:55.854674 6276 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 11:40:55.854687 6276 factory.go:656] Stopping watch factory\\\\nI0122 11:40:55.854697 6276 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:55.854712 6276 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854735 6276 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 11:40:55.854740 6276 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0122 11:40:55.854749 6276 handler.go:208] Removed *v1.Pod event handler 6\\\\nF0122 11:40:55.854804 6276 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.284707 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.297132 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.297187 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.297203 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.297223 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.297246 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.303527 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.323989 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8470cb5-cfaf-4760-8c07-ce375052950f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.324053 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8470cb5-cfaf-4760-8c07-ce375052950f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.324100 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9smw\" (UniqueName: \"kubernetes.io/projected/a8470cb5-cfaf-4760-8c07-ce375052950f-kube-api-access-t9smw\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.324167 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8470cb5-cfaf-4760-8c07-ce375052950f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.324599 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8470cb5-cfaf-4760-8c07-ce375052950f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.325164 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8470cb5-cfaf-4760-8c07-ce375052950f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.328707 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.336314 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8470cb5-cfaf-4760-8c07-ce375052950f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.352337 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.357928 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9smw\" (UniqueName: \"kubernetes.io/projected/a8470cb5-cfaf-4760-8c07-ce375052950f-kube-api-access-t9smw\") pod \"ovnkube-control-plane-749d76644c-x5vd5\" (UID: \"a8470cb5-cfaf-4760-8c07-ce375052950f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.376039 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.387374 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.400012 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.400053 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.400061 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.400074 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.400090 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.402190 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" Jan 22 11:40:57 crc kubenswrapper[4874]: W0122 11:40:57.411988 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8470cb5_cfaf_4760_8c07_ce375052950f.slice/crio-14cea2c3844143377a5f42b7666cebbe3d097f531368e488b35f8d214024a66f WatchSource:0}: Error finding container 14cea2c3844143377a5f42b7666cebbe3d097f531368e488b35f8d214024a66f: Status 404 returned error can't find the container with id 14cea2c3844143377a5f42b7666cebbe3d097f531368e488b35f8d214024a66f Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.452340 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.463946 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.474002 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.483092 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.492342 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.503005 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.503041 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.503050 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.503064 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.503075 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.504262 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.521347 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.532730 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.542907 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.556056 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.567591 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.585137 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:54Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298122 6154 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298314 6154 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 11:40:54.299178 6154 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:40:54.299220 6154 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 11:40:54.299250 6154 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:40:54.299337 6154 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 11:40:54.299719 6154 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:40:54.299899 6154 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 11:40:54.299923 6154 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:54.299953 6154 factory.go:656] Stopping watch factory\\\\nI0122 11:40:54.299985 6154 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:54.300031 6154 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:54.300052 6154 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:56Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI0122 11:40:55.854329 6276 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:55.854364 6276 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 11:40:55.854411 6276 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 11:40:55.854421 6276 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:55.854565 6276 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854637 6276 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0122 11:40:55.854653 6276 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:55.854670 6276 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 11:40:55.854674 6276 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 11:40:55.854687 6276 factory.go:656] Stopping watch factory\\\\nI0122 11:40:55.854697 6276 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:55.854712 6276 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854735 6276 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 11:40:55.854740 6276 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0122 11:40:55.854749 6276 handler.go:208] Removed *v1.Pod event handler 6\\\\nF0122 11:40:55.854804 6276 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.596552 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.605283 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.605320 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.605328 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.605346 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.605356 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.609011 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.621810 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.634905 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.647959 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.675993 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 21:19:17.694919958 +0000 UTC Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.708173 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.708200 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.708208 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.708220 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.708228 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.746776 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.746819 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.746831 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.746848 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.746859 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: E0122 11:40:57.758091 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.761454 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.761500 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.761512 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.761530 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.761541 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: E0122 11:40:57.774575 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.782836 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.782895 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.782910 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.782931 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.782947 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: E0122 11:40:57.793985 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.797141 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.797177 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.797186 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.797201 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.797213 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: E0122 11:40:57.808712 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.812783 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.812821 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.812834 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.812849 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.812859 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: E0122 11:40:57.825735 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: E0122 11:40:57.825911 4874 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.827346 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.827384 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.827424 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.827447 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.827460 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.929418 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.929480 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.929491 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.929532 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.929544 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:57Z","lastTransitionTime":"2026-01-22T11:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.931161 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" event={"ID":"a8470cb5-cfaf-4760-8c07-ce375052950f","Type":"ContainerStarted","Data":"abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.931263 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" event={"ID":"a8470cb5-cfaf-4760-8c07-ce375052950f","Type":"ContainerStarted","Data":"851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.931287 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" event={"ID":"a8470cb5-cfaf-4760-8c07-ce375052950f","Type":"ContainerStarted","Data":"14cea2c3844143377a5f42b7666cebbe3d097f531368e488b35f8d214024a66f"} Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.933074 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/1.log" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.936275 4874 scope.go:117] "RemoveContainer" containerID="6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9" Jan 22 11:40:57 crc kubenswrapper[4874]: E0122 11:40:57.936572 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.951676 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.964638 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.977632 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:57 crc kubenswrapper[4874]: I0122 11:40:57.988795 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:57Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.006482 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.019905 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.030317 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.031204 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.031299 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.031309 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.031331 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.031340 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:58Z","lastTransitionTime":"2026-01-22T11:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.042133 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.057067 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.081068 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72124241a469696f769c8a49bed7fe7008d988d3139c7894e11129e25247d0c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:54Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298122 6154 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:40:54.298314 6154 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 11:40:54.299178 6154 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:40:54.299220 6154 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 11:40:54.299250 6154 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:40:54.299337 6154 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 11:40:54.299719 6154 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:40:54.299899 6154 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 11:40:54.299923 6154 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:54.299953 6154 factory.go:656] Stopping watch factory\\\\nI0122 11:40:54.299985 6154 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:54.300031 6154 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:54.300052 6154 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:56Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI0122 11:40:55.854329 6276 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:55.854364 6276 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 11:40:55.854411 6276 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 11:40:55.854421 6276 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:55.854565 6276 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854637 6276 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0122 11:40:55.854653 6276 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:55.854670 6276 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 11:40:55.854674 6276 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 11:40:55.854687 6276 factory.go:656] Stopping watch factory\\\\nI0122 11:40:55.854697 6276 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:55.854712 6276 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854735 6276 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 11:40:55.854740 6276 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0122 11:40:55.854749 6276 handler.go:208] Removed *v1.Pod event handler 6\\\\nF0122 11:40:55.854804 6276 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.092557 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.102963 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.115239 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.132875 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.134080 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.134140 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.134163 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.134188 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.134206 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:58Z","lastTransitionTime":"2026-01-22T11:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.146329 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.162981 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.191437 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.218229 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.234699 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.236433 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.236573 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.236660 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.236741 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.236825 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:58Z","lastTransitionTime":"2026-01-22T11:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.250923 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.266990 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.285733 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.305973 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:56Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI0122 11:40:55.854329 6276 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:55.854364 6276 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 11:40:55.854411 6276 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 11:40:55.854421 6276 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:55.854565 6276 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854637 6276 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0122 11:40:55.854653 6276 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:55.854670 6276 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 11:40:55.854674 6276 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 11:40:55.854687 6276 factory.go:656] Stopping watch factory\\\\nI0122 11:40:55.854697 6276 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:55.854712 6276 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854735 6276 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 11:40:55.854740 6276 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0122 11:40:55.854749 6276 handler.go:208] Removed *v1.Pod event handler 6\\\\nF0122 11:40:55.854804 6276 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.321102 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.334420 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.339037 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.339073 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.339088 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.339106 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.339118 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:58Z","lastTransitionTime":"2026-01-22T11:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.350154 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.364933 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.386759 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.427922 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.440948 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.440996 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.441005 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.441020 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.441031 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:58Z","lastTransitionTime":"2026-01-22T11:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.466933 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.507333 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.527225 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-lr2vd"] Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.527858 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:40:58 crc kubenswrapper[4874]: E0122 11:40:58.527924 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.543652 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.543912 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.544020 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.544133 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.544214 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:58Z","lastTransitionTime":"2026-01-22T11:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.546324 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.593687 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.633467 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.638310 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.638356 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmzdj\" (UniqueName: \"kubernetes.io/projected/5451fbab-ebad-42e7-bb80-f94bad10d571-kube-api-access-hmzdj\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.647638 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.647698 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.647721 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.647754 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.647776 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:58Z","lastTransitionTime":"2026-01-22T11:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.672689 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.676968 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 10:28:53.114753414 +0000 UTC Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.712159 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.715503 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.715503 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:40:58 crc kubenswrapper[4874]: E0122 11:40:58.715721 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:40:58 crc kubenswrapper[4874]: E0122 11:40:58.715782 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.715520 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:40:58 crc kubenswrapper[4874]: E0122 11:40:58.715968 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.739174 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.739210 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmzdj\" (UniqueName: \"kubernetes.io/projected/5451fbab-ebad-42e7-bb80-f94bad10d571-kube-api-access-hmzdj\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:40:58 crc kubenswrapper[4874]: E0122 11:40:58.739486 4874 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:40:58 crc kubenswrapper[4874]: E0122 11:40:58.739717 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs podName:5451fbab-ebad-42e7-bb80-f94bad10d571 nodeName:}" failed. No retries permitted until 2026-01-22 11:40:59.239615069 +0000 UTC m=+33.084686179 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs") pod "network-metrics-daemon-lr2vd" (UID: "5451fbab-ebad-42e7-bb80-f94bad10d571") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.746840 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.750482 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.750511 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.750520 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.750534 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.750543 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:58Z","lastTransitionTime":"2026-01-22T11:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.776444 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmzdj\" (UniqueName: \"kubernetes.io/projected/5451fbab-ebad-42e7-bb80-f94bad10d571-kube-api-access-hmzdj\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.812623 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.853225 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.853273 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.853290 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.853309 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.853322 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:58Z","lastTransitionTime":"2026-01-22T11:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.853821 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.901818 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:56Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI0122 11:40:55.854329 6276 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:55.854364 6276 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 11:40:55.854411 6276 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 11:40:55.854421 6276 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:55.854565 6276 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854637 6276 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0122 11:40:55.854653 6276 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:55.854670 6276 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 11:40:55.854674 6276 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 11:40:55.854687 6276 factory.go:656] Stopping watch factory\\\\nI0122 11:40:55.854697 6276 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:55.854712 6276 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854735 6276 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 11:40:55.854740 6276 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0122 11:40:55.854749 6276 handler.go:208] Removed *v1.Pod event handler 6\\\\nF0122 11:40:55.854804 6276 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.927803 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.955234 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.955271 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.955280 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.955295 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.955304 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:58Z","lastTransitionTime":"2026-01-22T11:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:58 crc kubenswrapper[4874]: I0122 11:40:58.966828 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:58Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.010063 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:59Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.044291 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:59Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.058020 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.058075 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.058088 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.058104 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.058116 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:59Z","lastTransitionTime":"2026-01-22T11:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.086904 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:59Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.125424 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:59Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.160761 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.160817 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.160830 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.160845 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.160855 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:59Z","lastTransitionTime":"2026-01-22T11:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.167133 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:59Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.203900 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:59Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.244255 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:40:59 crc kubenswrapper[4874]: E0122 11:40:59.244456 4874 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:40:59 crc kubenswrapper[4874]: E0122 11:40:59.244513 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs podName:5451fbab-ebad-42e7-bb80-f94bad10d571 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:00.244496279 +0000 UTC m=+34.089567349 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs") pod "network-metrics-daemon-lr2vd" (UID: "5451fbab-ebad-42e7-bb80-f94bad10d571") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.248623 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:40:59Z is after 2025-08-24T17:21:41Z" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.262992 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.263040 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.263052 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.263067 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.263076 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:59Z","lastTransitionTime":"2026-01-22T11:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.366163 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.366220 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.366232 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.366245 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.366256 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:59Z","lastTransitionTime":"2026-01-22T11:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.468688 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.468746 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.468756 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.468771 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.468781 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:59Z","lastTransitionTime":"2026-01-22T11:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.571668 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.571720 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.571735 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.571754 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.571770 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:59Z","lastTransitionTime":"2026-01-22T11:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.674256 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.674323 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.674342 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.674366 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.674382 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:59Z","lastTransitionTime":"2026-01-22T11:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.677584 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 07:41:10.586831957 +0000 UTC Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.715473 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:40:59 crc kubenswrapper[4874]: E0122 11:40:59.715655 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.777044 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.777110 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.777126 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.777147 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.777164 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:59Z","lastTransitionTime":"2026-01-22T11:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.880074 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.880112 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.880126 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.880220 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.880273 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:59Z","lastTransitionTime":"2026-01-22T11:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.983655 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.983709 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.983723 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.983737 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:40:59 crc kubenswrapper[4874]: I0122 11:40:59.983747 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:40:59Z","lastTransitionTime":"2026-01-22T11:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.086434 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.086482 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.086495 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.086513 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.086525 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:00Z","lastTransitionTime":"2026-01-22T11:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.190187 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.190249 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.190265 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.190288 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.190305 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:00Z","lastTransitionTime":"2026-01-22T11:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.254663 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.254838 4874 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.254910 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs podName:5451fbab-ebad-42e7-bb80-f94bad10d571 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:02.254886969 +0000 UTC m=+36.099958079 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs") pod "network-metrics-daemon-lr2vd" (UID: "5451fbab-ebad-42e7-bb80-f94bad10d571") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.292885 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.292961 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.292984 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.293013 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.293036 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:00Z","lastTransitionTime":"2026-01-22T11:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.396554 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.396618 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.396636 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.396660 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.396679 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:00Z","lastTransitionTime":"2026-01-22T11:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.499503 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.499547 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.499558 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.499574 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.499584 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:00Z","lastTransitionTime":"2026-01-22T11:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.557928 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.558095 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:41:16.558066673 +0000 UTC m=+50.403137783 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.558141 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.558192 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.558229 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.558477 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.558513 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.558533 4874 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.558581 4874 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.558587 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:16.558571799 +0000 UTC m=+50.403642909 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.558668 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:16.558649762 +0000 UTC m=+50.403720862 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.558720 4874 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.558762 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:16.558747195 +0000 UTC m=+50.403818305 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.602734 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.602772 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.602787 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.602805 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.602816 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:00Z","lastTransitionTime":"2026-01-22T11:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.660039 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.660366 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.660461 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.660490 4874 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.660588 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:16.660555083 +0000 UTC m=+50.505626193 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.677894 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 16:14:02.824167179 +0000 UTC Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.706084 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.706160 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.706185 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.706261 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.706286 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:00Z","lastTransitionTime":"2026-01-22T11:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.716107 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.716114 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.716330 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.716316 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.716473 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:00 crc kubenswrapper[4874]: E0122 11:41:00.716556 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.809478 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.809543 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.809562 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.809588 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.809605 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:00Z","lastTransitionTime":"2026-01-22T11:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.913219 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.913298 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.913321 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.913354 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:00 crc kubenswrapper[4874]: I0122 11:41:00.913378 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:00Z","lastTransitionTime":"2026-01-22T11:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.016312 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.016361 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.016375 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.016414 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.016430 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:01Z","lastTransitionTime":"2026-01-22T11:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.118821 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.118882 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.118900 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.118924 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.118940 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:01Z","lastTransitionTime":"2026-01-22T11:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.221657 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.221707 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.221720 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.221742 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.221755 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:01Z","lastTransitionTime":"2026-01-22T11:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.324282 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.324346 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.324359 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.324378 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.324449 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:01Z","lastTransitionTime":"2026-01-22T11:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.426940 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.426983 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.427018 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.427038 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.427049 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:01Z","lastTransitionTime":"2026-01-22T11:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.530373 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.530482 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.530505 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.530538 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.530560 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:01Z","lastTransitionTime":"2026-01-22T11:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.634070 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.634127 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.634140 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.634159 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.634173 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:01Z","lastTransitionTime":"2026-01-22T11:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.677982 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 22:39:45.492806958 +0000 UTC Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.715952 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:01 crc kubenswrapper[4874]: E0122 11:41:01.716213 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.736686 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.737055 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.737554 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.737774 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.738279 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:01Z","lastTransitionTime":"2026-01-22T11:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.840903 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.840972 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.840991 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.841014 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.841033 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:01Z","lastTransitionTime":"2026-01-22T11:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.944037 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.944114 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.944139 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.944174 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:01 crc kubenswrapper[4874]: I0122 11:41:01.944246 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:01Z","lastTransitionTime":"2026-01-22T11:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.047080 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.047149 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.047176 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.047206 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.047226 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:02Z","lastTransitionTime":"2026-01-22T11:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.151038 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.151099 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.151121 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.151149 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.151174 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:02Z","lastTransitionTime":"2026-01-22T11:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.254955 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.255008 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.255020 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.255038 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.255050 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:02Z","lastTransitionTime":"2026-01-22T11:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.278057 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:02 crc kubenswrapper[4874]: E0122 11:41:02.278278 4874 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:41:02 crc kubenswrapper[4874]: E0122 11:41:02.278369 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs podName:5451fbab-ebad-42e7-bb80-f94bad10d571 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:06.278351845 +0000 UTC m=+40.123422915 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs") pod "network-metrics-daemon-lr2vd" (UID: "5451fbab-ebad-42e7-bb80-f94bad10d571") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.358115 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.358360 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.358462 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.358595 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.358722 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:02Z","lastTransitionTime":"2026-01-22T11:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.462042 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.462100 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.462118 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.462142 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.462159 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:02Z","lastTransitionTime":"2026-01-22T11:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.564896 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.565190 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.565288 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.565383 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.565504 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:02Z","lastTransitionTime":"2026-01-22T11:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.668805 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.669117 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.669253 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.669375 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.669528 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:02Z","lastTransitionTime":"2026-01-22T11:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.679162 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 02:05:55.943411946 +0000 UTC Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.715499 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.715523 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:02 crc kubenswrapper[4874]: E0122 11:41:02.715614 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:02 crc kubenswrapper[4874]: E0122 11:41:02.715691 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.715775 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:02 crc kubenswrapper[4874]: E0122 11:41:02.715840 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.771987 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.772023 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.772033 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.772047 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.772056 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:02Z","lastTransitionTime":"2026-01-22T11:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.874265 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.874318 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.874330 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.874342 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.874351 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:02Z","lastTransitionTime":"2026-01-22T11:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.976574 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.976835 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.976924 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.977047 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:02 crc kubenswrapper[4874]: I0122 11:41:02.977141 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:02Z","lastTransitionTime":"2026-01-22T11:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.079530 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.079569 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.079582 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.079600 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.079612 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:03Z","lastTransitionTime":"2026-01-22T11:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.182323 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.182377 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.182419 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.182444 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.182461 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:03Z","lastTransitionTime":"2026-01-22T11:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.286363 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.286457 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.286476 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.286504 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.286522 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:03Z","lastTransitionTime":"2026-01-22T11:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.389181 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.389205 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.389212 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.389226 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.389236 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:03Z","lastTransitionTime":"2026-01-22T11:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.491314 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.491623 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.491695 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.491774 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.491845 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:03Z","lastTransitionTime":"2026-01-22T11:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.593783 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.593815 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.593824 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.593837 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.593846 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:03Z","lastTransitionTime":"2026-01-22T11:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.679451 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 09:03:19.94719951 +0000 UTC Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.696368 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.696424 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.696439 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.696460 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.696476 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:03Z","lastTransitionTime":"2026-01-22T11:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.715903 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:03 crc kubenswrapper[4874]: E0122 11:41:03.716067 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.798954 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.799191 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.799271 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.799347 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.799475 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:03Z","lastTransitionTime":"2026-01-22T11:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.902467 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.902538 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.902563 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.902592 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:03 crc kubenswrapper[4874]: I0122 11:41:03.902614 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:03Z","lastTransitionTime":"2026-01-22T11:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.006152 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.006228 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.006251 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.006280 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.006301 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:04Z","lastTransitionTime":"2026-01-22T11:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.109604 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.109658 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.109676 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.109701 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.109720 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:04Z","lastTransitionTime":"2026-01-22T11:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.212991 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.213055 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.213072 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.213097 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.213113 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:04Z","lastTransitionTime":"2026-01-22T11:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.316487 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.316554 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.316572 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.316595 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.316612 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:04Z","lastTransitionTime":"2026-01-22T11:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.419381 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.419436 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.419448 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.419464 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.419474 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:04Z","lastTransitionTime":"2026-01-22T11:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.522992 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.523081 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.523119 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.523148 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.523170 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:04Z","lastTransitionTime":"2026-01-22T11:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.626239 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.626291 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.626299 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.626314 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.626333 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:04Z","lastTransitionTime":"2026-01-22T11:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.680173 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 00:45:23.711531321 +0000 UTC Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.715731 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.715760 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.715889 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:04 crc kubenswrapper[4874]: E0122 11:41:04.715999 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:04 crc kubenswrapper[4874]: E0122 11:41:04.716159 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:04 crc kubenswrapper[4874]: E0122 11:41:04.716329 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.728240 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.728271 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.728281 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.728293 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.728303 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:04Z","lastTransitionTime":"2026-01-22T11:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.830529 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.830588 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.830605 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.830625 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.830640 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:04Z","lastTransitionTime":"2026-01-22T11:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.932627 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.932966 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.933082 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.933193 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:04 crc kubenswrapper[4874]: I0122 11:41:04.933294 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:04Z","lastTransitionTime":"2026-01-22T11:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.036575 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.036617 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.036627 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.036642 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.036654 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:05Z","lastTransitionTime":"2026-01-22T11:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.139927 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.139998 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.140023 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.140048 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.140065 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:05Z","lastTransitionTime":"2026-01-22T11:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.243073 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.243601 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.243766 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.243953 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.244139 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:05Z","lastTransitionTime":"2026-01-22T11:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.347284 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.347323 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.347332 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.347349 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.347360 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:05Z","lastTransitionTime":"2026-01-22T11:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.449592 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.449639 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.449652 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.449671 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.449681 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:05Z","lastTransitionTime":"2026-01-22T11:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.551310 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.551371 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.551389 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.551435 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.551451 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:05Z","lastTransitionTime":"2026-01-22T11:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.654136 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.654184 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.654197 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.654213 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.654223 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:05Z","lastTransitionTime":"2026-01-22T11:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.680605 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 15:31:19.570688528 +0000 UTC Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.715962 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:05 crc kubenswrapper[4874]: E0122 11:41:05.716090 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.756673 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.756707 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.756716 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.756732 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.756742 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:05Z","lastTransitionTime":"2026-01-22T11:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.859231 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.859535 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.859620 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.859696 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.859803 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:05Z","lastTransitionTime":"2026-01-22T11:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.965476 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.965523 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.965539 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.965555 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:05 crc kubenswrapper[4874]: I0122 11:41:05.965565 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:05Z","lastTransitionTime":"2026-01-22T11:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.068161 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.068374 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.068485 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.068606 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.068726 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:06Z","lastTransitionTime":"2026-01-22T11:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.172247 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.172304 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.172320 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.172343 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.172362 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:06Z","lastTransitionTime":"2026-01-22T11:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.275827 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.275904 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.275927 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.275956 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.275977 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:06Z","lastTransitionTime":"2026-01-22T11:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.322571 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:06 crc kubenswrapper[4874]: E0122 11:41:06.322786 4874 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:41:06 crc kubenswrapper[4874]: E0122 11:41:06.322857 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs podName:5451fbab-ebad-42e7-bb80-f94bad10d571 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:14.322839408 +0000 UTC m=+48.167910489 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs") pod "network-metrics-daemon-lr2vd" (UID: "5451fbab-ebad-42e7-bb80-f94bad10d571") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.378111 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.378190 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.378215 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.378239 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.378260 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:06Z","lastTransitionTime":"2026-01-22T11:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.480688 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.480722 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.480730 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.480745 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.480754 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:06Z","lastTransitionTime":"2026-01-22T11:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.583713 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.583776 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.583793 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.583861 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.583880 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:06Z","lastTransitionTime":"2026-01-22T11:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.681663 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 15:59:34.327535777 +0000 UTC Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.687019 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.687070 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.687082 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.687096 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.687106 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:06Z","lastTransitionTime":"2026-01-22T11:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.715826 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:06 crc kubenswrapper[4874]: E0122 11:41:06.715961 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.715838 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.716011 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:06 crc kubenswrapper[4874]: E0122 11:41:06.716057 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:06 crc kubenswrapper[4874]: E0122 11:41:06.716165 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.728323 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.744098 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.757230 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.772186 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.789889 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.789922 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.789930 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.789944 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.789954 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:06Z","lastTransitionTime":"2026-01-22T11:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.792672 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.828252 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.846452 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.865618 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.880106 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.892483 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.892541 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.892558 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.892581 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.892599 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:06Z","lastTransitionTime":"2026-01-22T11:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.901687 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.916989 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.934414 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:56Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI0122 11:40:55.854329 6276 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:55.854364 6276 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 11:40:55.854411 6276 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 11:40:55.854421 6276 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:55.854565 6276 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854637 6276 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0122 11:40:55.854653 6276 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:55.854670 6276 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 11:40:55.854674 6276 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 11:40:55.854687 6276 factory.go:656] Stopping watch factory\\\\nI0122 11:40:55.854697 6276 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:55.854712 6276 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854735 6276 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 11:40:55.854740 6276 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0122 11:40:55.854749 6276 handler.go:208] Removed *v1.Pod event handler 6\\\\nF0122 11:40:55.854804 6276 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.946211 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.984460 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.994949 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.995003 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.995017 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.995035 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:06 crc kubenswrapper[4874]: I0122 11:41:06.995047 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:06Z","lastTransitionTime":"2026-01-22T11:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.000994 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:06Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.023639 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:07Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.038797 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:07Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.098142 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.098211 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.098232 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.098261 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.098283 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.201492 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.201539 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.201551 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.201570 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.201581 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.304908 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.304966 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.304986 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.305015 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.305038 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.408132 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.408171 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.408184 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.408202 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.408217 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.511214 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.511656 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.511887 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.512146 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.512346 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.615657 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.615969 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.616148 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.616297 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.616534 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.682148 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 21:37:06.881433082 +0000 UTC Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.715896 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:07 crc kubenswrapper[4874]: E0122 11:41:07.716102 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.720678 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.720772 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.720798 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.720858 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.720888 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.823474 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.823528 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.823546 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.823569 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.823588 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.873320 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.873353 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.873364 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.873378 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.873389 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: E0122 11:41:07.894054 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:07Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.898972 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.899014 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.899024 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.899041 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.899053 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: E0122 11:41:07.913738 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:07Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.919168 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.919222 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.919238 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.919263 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.919282 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: E0122 11:41:07.938455 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:07Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.943336 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.943516 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.943539 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.943558 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.943574 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: E0122 11:41:07.961888 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:07Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.967283 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.967354 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.967371 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.967424 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.967444 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:07 crc kubenswrapper[4874]: E0122 11:41:07.987285 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:07Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:07 crc kubenswrapper[4874]: E0122 11:41:07.987553 4874 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.991153 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.991244 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.991262 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.991287 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:07 crc kubenswrapper[4874]: I0122 11:41:07.991304 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:07Z","lastTransitionTime":"2026-01-22T11:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.094057 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.094365 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.094662 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.095083 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.095623 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:08Z","lastTransitionTime":"2026-01-22T11:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.198730 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.199082 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.199362 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.199599 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.199736 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:08Z","lastTransitionTime":"2026-01-22T11:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.302515 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.302564 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.302586 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.302613 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.302635 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:08Z","lastTransitionTime":"2026-01-22T11:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.405791 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.405837 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.405853 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.405875 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.405891 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:08Z","lastTransitionTime":"2026-01-22T11:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.509254 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.509339 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.509363 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.509386 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.509428 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:08Z","lastTransitionTime":"2026-01-22T11:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.611932 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.611976 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.611993 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.612014 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.612031 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:08Z","lastTransitionTime":"2026-01-22T11:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.682950 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 11:32:20.683985974 +0000 UTC Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.715612 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.715658 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.715897 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:08 crc kubenswrapper[4874]: E0122 11:41:08.715934 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.715954 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.715995 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.716019 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.716037 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:08Z","lastTransitionTime":"2026-01-22T11:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:08 crc kubenswrapper[4874]: E0122 11:41:08.716059 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.716558 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:08 crc kubenswrapper[4874]: E0122 11:41:08.716700 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.819232 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.819285 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.819296 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.819313 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.819324 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:08Z","lastTransitionTime":"2026-01-22T11:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.922465 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.922545 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.922564 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.922590 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:08 crc kubenswrapper[4874]: I0122 11:41:08.922608 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:08Z","lastTransitionTime":"2026-01-22T11:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.026379 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.026468 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.026486 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.026510 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.026527 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:09Z","lastTransitionTime":"2026-01-22T11:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.128952 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.129024 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.129059 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.129108 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.129146 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:09Z","lastTransitionTime":"2026-01-22T11:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.232488 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.232549 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.232565 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.232590 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.232607 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:09Z","lastTransitionTime":"2026-01-22T11:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.334907 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.334969 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.334984 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.335000 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.335011 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:09Z","lastTransitionTime":"2026-01-22T11:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.437059 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.437097 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.437109 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.437127 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.437140 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:09Z","lastTransitionTime":"2026-01-22T11:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.539922 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.539983 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.539999 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.540023 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.540040 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:09Z","lastTransitionTime":"2026-01-22T11:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.643961 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.644022 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.644044 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.644076 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.644099 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:09Z","lastTransitionTime":"2026-01-22T11:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.683972 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 06:35:35.304586327 +0000 UTC Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.715510 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:09 crc kubenswrapper[4874]: E0122 11:41:09.715765 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.746743 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.746806 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.746829 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.746854 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.746873 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:09Z","lastTransitionTime":"2026-01-22T11:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.850650 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.850696 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.850708 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.850727 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.850742 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:09Z","lastTransitionTime":"2026-01-22T11:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.952905 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.952943 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.952953 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.952967 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:09 crc kubenswrapper[4874]: I0122 11:41:09.952990 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:09Z","lastTransitionTime":"2026-01-22T11:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.055095 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.055146 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.055157 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.055175 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.055185 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:10Z","lastTransitionTime":"2026-01-22T11:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.157828 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.157864 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.157876 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.157891 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.157901 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:10Z","lastTransitionTime":"2026-01-22T11:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.260795 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.260893 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.260912 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.261377 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.261625 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:10Z","lastTransitionTime":"2026-01-22T11:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.364706 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.364756 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.364770 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.364792 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.364807 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:10Z","lastTransitionTime":"2026-01-22T11:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.467905 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.468200 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.468479 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.468517 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.468530 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:10Z","lastTransitionTime":"2026-01-22T11:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.570883 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.570932 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.570950 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.570972 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.570991 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:10Z","lastTransitionTime":"2026-01-22T11:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.674041 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.674094 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.674113 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.674137 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.674154 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:10Z","lastTransitionTime":"2026-01-22T11:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.686102 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 12:29:11.94413328 +0000 UTC Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.715886 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.715878 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.715915 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:10 crc kubenswrapper[4874]: E0122 11:41:10.716027 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.716725 4874 scope.go:117] "RemoveContainer" containerID="6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9" Jan 22 11:41:10 crc kubenswrapper[4874]: E0122 11:41:10.716719 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:10 crc kubenswrapper[4874]: E0122 11:41:10.716970 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.778258 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.778318 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.778336 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.778362 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.778380 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:10Z","lastTransitionTime":"2026-01-22T11:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.881472 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.882002 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.882021 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.883216 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.884043 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:10Z","lastTransitionTime":"2026-01-22T11:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.986872 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.986901 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.986911 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.986923 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:10 crc kubenswrapper[4874]: I0122 11:41:10.986933 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:10Z","lastTransitionTime":"2026-01-22T11:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.003288 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/1.log" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.005512 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337"} Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.006673 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.016602 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.027578 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.039903 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.050929 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.062212 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.077488 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.089488 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.089798 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.089830 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.089842 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.089858 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.089871 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:11Z","lastTransitionTime":"2026-01-22T11:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.107949 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.132853 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.166977 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:56Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI0122 11:40:55.854329 6276 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:55.854364 6276 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 11:40:55.854411 6276 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 11:40:55.854421 6276 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:55.854565 6276 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854637 6276 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0122 11:40:55.854653 6276 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:55.854670 6276 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 11:40:55.854674 6276 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 11:40:55.854687 6276 factory.go:656] Stopping watch factory\\\\nI0122 11:40:55.854697 6276 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:55.854712 6276 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854735 6276 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 11:40:55.854740 6276 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0122 11:40:55.854749 6276 handler.go:208] Removed *v1.Pod event handler 6\\\\nF0122 11:40:55.854804 6276 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:41:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.188079 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.192014 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.192041 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.192051 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.192066 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.192075 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:11Z","lastTransitionTime":"2026-01-22T11:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.201342 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.212258 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.229637 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.252409 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.271471 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.284677 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:11Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.294156 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.294194 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.294203 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.294217 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.294225 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:11Z","lastTransitionTime":"2026-01-22T11:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.396274 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.396316 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.396328 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.396343 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.396355 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:11Z","lastTransitionTime":"2026-01-22T11:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.499210 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.499255 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.499271 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.499291 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.499308 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:11Z","lastTransitionTime":"2026-01-22T11:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.602034 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.602080 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.602091 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.602108 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.602121 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:11Z","lastTransitionTime":"2026-01-22T11:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.687454 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 11:32:38.566116847 +0000 UTC Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.710516 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.710615 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.710633 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.710656 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.710672 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:11Z","lastTransitionTime":"2026-01-22T11:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.715204 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:11 crc kubenswrapper[4874]: E0122 11:41:11.715384 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.813501 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.813569 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.813588 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.813613 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.813631 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:11Z","lastTransitionTime":"2026-01-22T11:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.917312 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.917369 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.917388 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.917461 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:11 crc kubenswrapper[4874]: I0122 11:41:11.917480 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:11Z","lastTransitionTime":"2026-01-22T11:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.012449 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/2.log" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.013450 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/1.log" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.018228 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337" exitCode=1 Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.018283 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337"} Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.018370 4874 scope.go:117] "RemoveContainer" containerID="6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.019334 4874 scope.go:117] "RemoveContainer" containerID="fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337" Jan 22 11:41:12 crc kubenswrapper[4874]: E0122 11:41:12.019677 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.019825 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.019895 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.019922 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.019954 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.019980 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:12Z","lastTransitionTime":"2026-01-22T11:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.064227 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.078469 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.095360 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.105866 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.120982 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.122669 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.122702 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.122710 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.122725 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.122736 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:12Z","lastTransitionTime":"2026-01-22T11:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.136182 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.156509 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.170306 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.186567 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.201270 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.220668 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.224120 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.224154 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.224163 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.224177 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.224187 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:12Z","lastTransitionTime":"2026-01-22T11:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.235194 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.252807 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.264845 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.279442 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.292327 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.309292 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cef98ffee64a9c18ee659084940bd115933c3b805e846f3798a4af2a27cb0c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:40:56Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI0122 11:40:55.854329 6276 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:40:55.854364 6276 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 11:40:55.854411 6276 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 11:40:55.854421 6276 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 11:40:55.854565 6276 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854637 6276 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0122 11:40:55.854653 6276 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 11:40:55.854670 6276 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 11:40:55.854674 6276 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 11:40:55.854687 6276 factory.go:656] Stopping watch factory\\\\nI0122 11:40:55.854697 6276 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:40:55.854712 6276 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 11:40:55.854735 6276 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 11:40:55.854740 6276 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0122 11:40:55.854749 6276 handler.go:208] Removed *v1.Pod event handler 6\\\\nF0122 11:40:55.854804 6276 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:11Z\\\",\\\"message\\\":\\\"actory.go:160\\\\nI0122 11:41:11.510239 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510274 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510315 6499 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 11:41:11.516659 6499 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:41:11.516689 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:41:11.516719 6499 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:41:11.516760 6499 factory.go:656] Stopping watch factory\\\\nI0122 11:41:11.516776 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:41:11.529962 6499 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0122 11:41:11.529996 6499 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0122 11:41:11.530061 6499 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:41:11.530100 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 11:41:11.530194 6499 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:12Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.327526 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.327564 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.327573 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.327587 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.327596 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:12Z","lastTransitionTime":"2026-01-22T11:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.430348 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.430451 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.430476 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.430502 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.430518 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:12Z","lastTransitionTime":"2026-01-22T11:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.533053 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.533141 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.533163 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.533190 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.533207 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:12Z","lastTransitionTime":"2026-01-22T11:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.636269 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.636329 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.636345 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.636368 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.636438 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:12Z","lastTransitionTime":"2026-01-22T11:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.687868 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 05:44:03.324922232 +0000 UTC Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.715653 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.715668 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.715769 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:12 crc kubenswrapper[4874]: E0122 11:41:12.715850 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:12 crc kubenswrapper[4874]: E0122 11:41:12.715913 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:12 crc kubenswrapper[4874]: E0122 11:41:12.715937 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.739114 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.739158 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.739169 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.739187 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.739199 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:12Z","lastTransitionTime":"2026-01-22T11:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.840915 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.840958 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.840966 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.840982 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.840992 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:12Z","lastTransitionTime":"2026-01-22T11:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.944123 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.944195 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.944218 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.944249 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:12 crc kubenswrapper[4874]: I0122 11:41:12.944268 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:12Z","lastTransitionTime":"2026-01-22T11:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.025609 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/2.log" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.032290 4874 scope.go:117] "RemoveContainer" containerID="fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337" Jan 22 11:41:13 crc kubenswrapper[4874]: E0122 11:41:13.032652 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.048207 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.048275 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.048298 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.048330 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.048353 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:13Z","lastTransitionTime":"2026-01-22T11:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.056393 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.077712 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.110454 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:11Z\\\",\\\"message\\\":\\\"actory.go:160\\\\nI0122 11:41:11.510239 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510274 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510315 6499 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 11:41:11.516659 6499 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:41:11.516689 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:41:11.516719 6499 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:41:11.516760 6499 factory.go:656] Stopping watch factory\\\\nI0122 11:41:11.516776 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:41:11.529962 6499 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0122 11:41:11.529996 6499 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0122 11:41:11.530061 6499 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:41:11.530100 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 11:41:11.530194 6499 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.128795 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.145298 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.151771 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.151834 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.151852 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.151876 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.151894 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:13Z","lastTransitionTime":"2026-01-22T11:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.164674 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.178335 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.195610 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.211118 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.229934 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.241857 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.253981 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.254035 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.254048 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.254067 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.254080 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:13Z","lastTransitionTime":"2026-01-22T11:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.257515 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.275738 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.302993 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.316975 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.333802 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.349493 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:13Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.356208 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.356246 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.356259 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.356275 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.356286 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:13Z","lastTransitionTime":"2026-01-22T11:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.459182 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.459259 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.459282 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.459314 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.459337 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:13Z","lastTransitionTime":"2026-01-22T11:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.563463 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.563507 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.563550 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.563576 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.563592 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:13Z","lastTransitionTime":"2026-01-22T11:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.666369 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.666434 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.666444 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.666457 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.666467 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:13Z","lastTransitionTime":"2026-01-22T11:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.688690 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 11:48:04.703006194 +0000 UTC Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.715254 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:13 crc kubenswrapper[4874]: E0122 11:41:13.715428 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.770028 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.770107 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.770130 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.770159 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.770182 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:13Z","lastTransitionTime":"2026-01-22T11:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.873503 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.873553 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.873564 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.873578 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.873587 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:13Z","lastTransitionTime":"2026-01-22T11:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.976714 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.976778 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.976796 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.976820 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:13 crc kubenswrapper[4874]: I0122 11:41:13.976849 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:13Z","lastTransitionTime":"2026-01-22T11:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.079456 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.079518 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.079540 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.079573 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.079596 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:14Z","lastTransitionTime":"2026-01-22T11:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.182818 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.182887 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.182911 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.182941 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.182963 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:14Z","lastTransitionTime":"2026-01-22T11:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.285944 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.285995 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.286011 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.286033 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.286061 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:14Z","lastTransitionTime":"2026-01-22T11:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.387882 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.387934 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.387942 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.387954 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.388001 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:14Z","lastTransitionTime":"2026-01-22T11:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.412447 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:14 crc kubenswrapper[4874]: E0122 11:41:14.412587 4874 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:41:14 crc kubenswrapper[4874]: E0122 11:41:14.412638 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs podName:5451fbab-ebad-42e7-bb80-f94bad10d571 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:30.412625253 +0000 UTC m=+64.257696313 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs") pod "network-metrics-daemon-lr2vd" (UID: "5451fbab-ebad-42e7-bb80-f94bad10d571") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.490262 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.490313 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.490329 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.490350 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.490366 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:14Z","lastTransitionTime":"2026-01-22T11:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.606853 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.606958 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.606984 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.607013 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.607035 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:14Z","lastTransitionTime":"2026-01-22T11:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.688986 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 15:45:43.122171471 +0000 UTC Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.710260 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.710322 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.710333 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.710347 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.710359 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:14Z","lastTransitionTime":"2026-01-22T11:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.715858 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.715939 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:14 crc kubenswrapper[4874]: E0122 11:41:14.715995 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.716044 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:14 crc kubenswrapper[4874]: E0122 11:41:14.716266 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:14 crc kubenswrapper[4874]: E0122 11:41:14.716497 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.813250 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.813320 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.813340 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.813362 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.813379 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:14Z","lastTransitionTime":"2026-01-22T11:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.916420 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.916470 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.916480 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.916499 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:14 crc kubenswrapper[4874]: I0122 11:41:14.916512 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:14Z","lastTransitionTime":"2026-01-22T11:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.019344 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.019376 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.019384 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.019425 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.019434 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:15Z","lastTransitionTime":"2026-01-22T11:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.122706 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.122769 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.122844 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.122869 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.122886 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:15Z","lastTransitionTime":"2026-01-22T11:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.225568 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.225654 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.225680 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.225710 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.225732 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:15Z","lastTransitionTime":"2026-01-22T11:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.240179 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.257118 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.261058 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.277706 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.296634 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.316765 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.328844 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.329117 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.329252 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.329460 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.329847 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:15Z","lastTransitionTime":"2026-01-22T11:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.349986 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.369360 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.385840 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.401129 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.423286 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.433468 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.433519 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.433533 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.433557 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.433574 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:15Z","lastTransitionTime":"2026-01-22T11:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.439071 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.463840 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:11Z\\\",\\\"message\\\":\\\"actory.go:160\\\\nI0122 11:41:11.510239 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510274 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510315 6499 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 11:41:11.516659 6499 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:41:11.516689 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:41:11.516719 6499 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:41:11.516760 6499 factory.go:656] Stopping watch factory\\\\nI0122 11:41:11.516776 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:41:11.529962 6499 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0122 11:41:11.529996 6499 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0122 11:41:11.530061 6499 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:41:11.530100 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 11:41:11.530194 6499 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.484804 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.502955 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.518877 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.535740 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.535779 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.535791 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.535806 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.535817 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:15Z","lastTransitionTime":"2026-01-22T11:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.540343 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.554514 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.565089 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:15Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.638988 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.639171 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.639279 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.639384 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.639533 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:15Z","lastTransitionTime":"2026-01-22T11:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.691693 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 20:04:11.154455496 +0000 UTC Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.716015 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:15 crc kubenswrapper[4874]: E0122 11:41:15.716156 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.742051 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.742085 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.742097 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.742114 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.742126 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:15Z","lastTransitionTime":"2026-01-22T11:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.844900 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.844975 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.844989 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.845007 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.845030 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:15Z","lastTransitionTime":"2026-01-22T11:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.948543 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.948595 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.948608 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.948626 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:15 crc kubenswrapper[4874]: I0122 11:41:15.948640 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:15Z","lastTransitionTime":"2026-01-22T11:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.050520 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.050558 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.050569 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.050586 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.050597 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:16Z","lastTransitionTime":"2026-01-22T11:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.154072 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.154144 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.154160 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.154180 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.154195 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:16Z","lastTransitionTime":"2026-01-22T11:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.257045 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.257087 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.257121 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.257139 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.257150 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:16Z","lastTransitionTime":"2026-01-22T11:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.360149 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.360206 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.360226 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.360249 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.360266 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:16Z","lastTransitionTime":"2026-01-22T11:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.463870 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.463936 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.463953 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.463985 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.464024 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:16Z","lastTransitionTime":"2026-01-22T11:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.567800 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.567875 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.567906 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.567938 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.567960 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:16Z","lastTransitionTime":"2026-01-22T11:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.637707 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.637896 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:41:48.637859987 +0000 UTC m=+82.482931127 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.638020 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.638084 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.638137 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.638278 4874 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.638348 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.638354 4874 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.638366 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:48.638342183 +0000 UTC m=+82.483413283 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.638380 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.638983 4874 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.639093 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:48.639061456 +0000 UTC m=+82.484132566 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.639200 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:48.63916747 +0000 UTC m=+82.484238590 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.670849 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.670910 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.670922 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.670940 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.670949 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:16Z","lastTransitionTime":"2026-01-22T11:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.693317 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 05:23:41.263561064 +0000 UTC Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.715144 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.715219 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.715273 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.715442 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.715500 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.715671 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.735903 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d542969f-3655-4a4e-8a4d-238cff44f86e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb2908cd58fbcf7ae2f0e4281b7c0ef1a68896ab514d9aa90f347f7346b479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9d6d3c847805c81649e7524a51e0d1a261d3c75d10c90e9e2d6d6a0723ff76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0b6dcbe2e50cd1aa3a3e2edbc5401888ebd6f99cbbf5329245bd4f61bf7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.738716 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.738899 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.738918 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.738933 4874 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:16 crc kubenswrapper[4874]: E0122 11:41:16.738986 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 11:41:48.738971854 +0000 UTC m=+82.584042944 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.759110 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.773855 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.773896 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.773905 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.773937 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.773947 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:16Z","lastTransitionTime":"2026-01-22T11:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.779739 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.799187 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.811834 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.824945 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.839994 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.854486 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.864808 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.876514 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.876543 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.876565 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.876579 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.876601 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:16Z","lastTransitionTime":"2026-01-22T11:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.880619 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.892983 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.910560 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.922076 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.933549 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.942385 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.958905 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.972214 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.978559 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.978586 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.978595 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.978609 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.978619 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:16Z","lastTransitionTime":"2026-01-22T11:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:16 crc kubenswrapper[4874]: I0122 11:41:16.992426 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:11Z\\\",\\\"message\\\":\\\"actory.go:160\\\\nI0122 11:41:11.510239 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510274 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510315 6499 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 11:41:11.516659 6499 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:41:11.516689 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:41:11.516719 6499 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:41:11.516760 6499 factory.go:656] Stopping watch factory\\\\nI0122 11:41:11.516776 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:41:11.529962 6499 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0122 11:41:11.529996 6499 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0122 11:41:11.530061 6499 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:41:11.530100 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 11:41:11.530194 6499 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:16Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.080750 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.080809 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.080831 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.080861 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.080882 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:17Z","lastTransitionTime":"2026-01-22T11:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.183859 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.183910 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.183927 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.183951 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.183967 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:17Z","lastTransitionTime":"2026-01-22T11:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.286496 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.286533 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.286545 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.286560 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.286570 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:17Z","lastTransitionTime":"2026-01-22T11:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.389229 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.389285 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.389301 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.389325 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.389343 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:17Z","lastTransitionTime":"2026-01-22T11:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.492739 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.493142 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.493302 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.493491 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.493690 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:17Z","lastTransitionTime":"2026-01-22T11:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.597012 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.597358 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.597653 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.597938 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.598167 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:17Z","lastTransitionTime":"2026-01-22T11:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.694465 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 00:30:45.555761875 +0000 UTC Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.701844 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.701915 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.701938 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.701969 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.701991 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:17Z","lastTransitionTime":"2026-01-22T11:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.716057 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:17 crc kubenswrapper[4874]: E0122 11:41:17.716226 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.805275 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.805329 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.805340 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.805360 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.805374 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:17Z","lastTransitionTime":"2026-01-22T11:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.908791 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.908842 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.908853 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.908870 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:17 crc kubenswrapper[4874]: I0122 11:41:17.908882 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:17Z","lastTransitionTime":"2026-01-22T11:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.011714 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.011746 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.011754 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.011769 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.011780 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.115033 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.115076 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.115088 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.115104 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.115115 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.217928 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.217986 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.218078 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.218109 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.218121 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.298445 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.298509 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.298526 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.298550 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.298567 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: E0122 11:41:18.319993 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:18Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.324837 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.324861 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.324870 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.324882 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.324890 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: E0122 11:41:18.344258 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:18Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.348984 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.349043 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.349067 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.349095 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.349122 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: E0122 11:41:18.366384 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:18Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.370772 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.370819 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.370898 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.370936 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.370952 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: E0122 11:41:18.389430 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:18Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.393042 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.393085 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.393097 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.393116 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.393450 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: E0122 11:41:18.411755 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:18Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:18 crc kubenswrapper[4874]: E0122 11:41:18.411870 4874 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.414089 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.414168 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.414186 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.414209 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.414227 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.517954 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.518074 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.518098 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.518128 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.518155 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.621507 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.621559 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.621569 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.621583 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.621593 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.695555 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 23:21:48.652884682 +0000 UTC Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.716031 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.716092 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:18 crc kubenswrapper[4874]: E0122 11:41:18.716197 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.716278 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:18 crc kubenswrapper[4874]: E0122 11:41:18.716542 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:18 crc kubenswrapper[4874]: E0122 11:41:18.716284 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.724819 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.724902 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.724924 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.724959 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.724980 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.828374 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.828461 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.828477 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.828502 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.828519 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.931004 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.931064 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.931082 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.931107 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:18 crc kubenswrapper[4874]: I0122 11:41:18.931132 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:18Z","lastTransitionTime":"2026-01-22T11:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.033995 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.034060 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.034080 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.034110 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.034132 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:19Z","lastTransitionTime":"2026-01-22T11:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.136824 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.136880 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.136897 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.136921 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.136952 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:19Z","lastTransitionTime":"2026-01-22T11:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.240787 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.240843 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.240862 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.240884 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.240904 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:19Z","lastTransitionTime":"2026-01-22T11:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.344227 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.344277 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.344289 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.344309 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.344320 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:19Z","lastTransitionTime":"2026-01-22T11:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.447849 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.447926 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.447944 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.448455 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.448513 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:19Z","lastTransitionTime":"2026-01-22T11:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.551690 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.551776 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.551804 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.551834 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.551856 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:19Z","lastTransitionTime":"2026-01-22T11:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.655126 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.655197 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.655219 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.655249 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.655274 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:19Z","lastTransitionTime":"2026-01-22T11:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.695831 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 15:16:59.969366602 +0000 UTC Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.715046 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:19 crc kubenswrapper[4874]: E0122 11:41:19.715189 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.760168 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.760233 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.760251 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.760276 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.760299 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:19Z","lastTransitionTime":"2026-01-22T11:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.863856 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.863923 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.863947 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.863978 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.863997 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:19Z","lastTransitionTime":"2026-01-22T11:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.966746 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.966829 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.966852 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.966882 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:19 crc kubenswrapper[4874]: I0122 11:41:19.966906 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:19Z","lastTransitionTime":"2026-01-22T11:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.070769 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.070836 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.070857 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.070881 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.070898 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:20Z","lastTransitionTime":"2026-01-22T11:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.173933 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.173991 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.174012 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.174036 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.174052 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:20Z","lastTransitionTime":"2026-01-22T11:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.277272 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.277347 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.277364 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.277393 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.277435 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:20Z","lastTransitionTime":"2026-01-22T11:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.381195 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.381241 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.381257 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.381278 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.381295 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:20Z","lastTransitionTime":"2026-01-22T11:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.484225 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.484292 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.484315 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.484344 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.484365 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:20Z","lastTransitionTime":"2026-01-22T11:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.587805 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.587871 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.587890 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.587914 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.587931 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:20Z","lastTransitionTime":"2026-01-22T11:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.690661 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.690735 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.690760 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.690788 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.690810 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:20Z","lastTransitionTime":"2026-01-22T11:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.695940 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 04:36:43.461722688 +0000 UTC Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.715326 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.715550 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.715431 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:20 crc kubenswrapper[4874]: E0122 11:41:20.715698 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:20 crc kubenswrapper[4874]: E0122 11:41:20.715810 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:20 crc kubenswrapper[4874]: E0122 11:41:20.715926 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.793747 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.793782 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.793794 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.793810 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.793822 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:20Z","lastTransitionTime":"2026-01-22T11:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.896356 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.896580 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.896604 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.896628 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:20 crc kubenswrapper[4874]: I0122 11:41:20.896648 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:20Z","lastTransitionTime":"2026-01-22T11:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.001883 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.001960 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.001979 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.002001 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.002020 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:21Z","lastTransitionTime":"2026-01-22T11:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.106173 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.106296 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.106327 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.106469 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.106573 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:21Z","lastTransitionTime":"2026-01-22T11:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.210225 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.210289 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.210306 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.210329 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.210346 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:21Z","lastTransitionTime":"2026-01-22T11:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.313473 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.313514 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.313526 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.313540 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.313549 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:21Z","lastTransitionTime":"2026-01-22T11:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.417216 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.417275 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.417292 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.417314 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.417330 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:21Z","lastTransitionTime":"2026-01-22T11:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.520778 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.520844 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.520861 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.520885 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.520902 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:21Z","lastTransitionTime":"2026-01-22T11:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.623271 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.623308 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.623319 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.623335 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.623371 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:21Z","lastTransitionTime":"2026-01-22T11:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.696079 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 07:18:49.615358566 +0000 UTC Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.715669 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:21 crc kubenswrapper[4874]: E0122 11:41:21.715789 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.725452 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.725511 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.725528 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.725554 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.725581 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:21Z","lastTransitionTime":"2026-01-22T11:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.828791 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.828873 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.828893 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.828916 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.828961 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:21Z","lastTransitionTime":"2026-01-22T11:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.932374 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.932503 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.932527 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.932553 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:21 crc kubenswrapper[4874]: I0122 11:41:21.932572 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:21Z","lastTransitionTime":"2026-01-22T11:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.036224 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.036285 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.036303 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.036327 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.036347 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:22Z","lastTransitionTime":"2026-01-22T11:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.139275 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.139338 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.139356 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.139383 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.139433 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:22Z","lastTransitionTime":"2026-01-22T11:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.242643 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.243109 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.243269 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.243482 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.243666 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:22Z","lastTransitionTime":"2026-01-22T11:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.346961 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.347010 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.347058 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.347077 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.347089 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:22Z","lastTransitionTime":"2026-01-22T11:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.449260 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.449322 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.449333 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.449351 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.449363 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:22Z","lastTransitionTime":"2026-01-22T11:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.552506 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.552566 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.552578 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.552593 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.552604 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:22Z","lastTransitionTime":"2026-01-22T11:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.654769 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.654819 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.654835 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.654855 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.654871 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:22Z","lastTransitionTime":"2026-01-22T11:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.696603 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 21:52:31.354151941 +0000 UTC Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.715170 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.715244 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.715313 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:22 crc kubenswrapper[4874]: E0122 11:41:22.715433 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:22 crc kubenswrapper[4874]: E0122 11:41:22.715575 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:22 crc kubenswrapper[4874]: E0122 11:41:22.715687 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.757630 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.757670 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.757684 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.757703 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.757719 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:22Z","lastTransitionTime":"2026-01-22T11:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.860691 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.860724 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.860732 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.860745 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.860754 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:22Z","lastTransitionTime":"2026-01-22T11:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.963119 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.963185 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.963202 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.963517 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:22 crc kubenswrapper[4874]: I0122 11:41:22.963540 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:22Z","lastTransitionTime":"2026-01-22T11:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.066877 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.066941 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.066963 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.066992 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.067013 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:23Z","lastTransitionTime":"2026-01-22T11:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.170207 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.170294 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.170319 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.170359 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.170384 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:23Z","lastTransitionTime":"2026-01-22T11:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.274048 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.274118 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.274137 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.274164 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.274187 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:23Z","lastTransitionTime":"2026-01-22T11:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.378011 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.378113 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.378141 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.378175 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.378216 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:23Z","lastTransitionTime":"2026-01-22T11:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.481249 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.481683 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.481841 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.481997 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.482134 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:23Z","lastTransitionTime":"2026-01-22T11:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.585874 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.585925 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.585937 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.585957 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.585971 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:23Z","lastTransitionTime":"2026-01-22T11:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.688465 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.688508 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.688520 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.688536 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.688547 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:23Z","lastTransitionTime":"2026-01-22T11:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.696765 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 20:12:27.323434594 +0000 UTC Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.715494 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:23 crc kubenswrapper[4874]: E0122 11:41:23.715681 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.790225 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.790286 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.790299 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.790311 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.790323 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:23Z","lastTransitionTime":"2026-01-22T11:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.893538 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.893586 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.893595 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.893608 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.893616 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:23Z","lastTransitionTime":"2026-01-22T11:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.996154 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.996185 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.996194 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.996206 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:23 crc kubenswrapper[4874]: I0122 11:41:23.996215 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:23Z","lastTransitionTime":"2026-01-22T11:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.097970 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.098033 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.098050 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.098074 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.098092 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:24Z","lastTransitionTime":"2026-01-22T11:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.201067 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.201131 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.201145 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.201180 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.201220 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:24Z","lastTransitionTime":"2026-01-22T11:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.304579 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.304663 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.304691 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.304718 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.304737 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:24Z","lastTransitionTime":"2026-01-22T11:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.407988 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.408082 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.408107 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.408135 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.408156 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:24Z","lastTransitionTime":"2026-01-22T11:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.511738 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.511801 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.511817 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.511842 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.511859 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:24Z","lastTransitionTime":"2026-01-22T11:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.614896 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.614967 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.614990 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.615015 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.615030 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:24Z","lastTransitionTime":"2026-01-22T11:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.697495 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 13:39:51.88347177 +0000 UTC Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.716164 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.716276 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:24 crc kubenswrapper[4874]: E0122 11:41:24.716387 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:24 crc kubenswrapper[4874]: E0122 11:41:24.716533 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.716679 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:24 crc kubenswrapper[4874]: E0122 11:41:24.716784 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.718377 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.718434 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.718463 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.718496 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.718509 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:24Z","lastTransitionTime":"2026-01-22T11:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.821820 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.821893 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.821913 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.821938 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.821979 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:24Z","lastTransitionTime":"2026-01-22T11:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.925075 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.925151 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.925181 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.925210 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:24 crc kubenswrapper[4874]: I0122 11:41:24.925232 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:24Z","lastTransitionTime":"2026-01-22T11:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.028793 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.028848 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.028863 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.028884 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.028898 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:25Z","lastTransitionTime":"2026-01-22T11:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.131547 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.131597 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.131613 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.131636 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.131652 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:25Z","lastTransitionTime":"2026-01-22T11:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.235386 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.235566 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.235594 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.235624 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.235648 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:25Z","lastTransitionTime":"2026-01-22T11:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.338326 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.338808 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.338961 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.339129 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.339255 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:25Z","lastTransitionTime":"2026-01-22T11:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.442735 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.442802 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.442820 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.442844 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.442861 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:25Z","lastTransitionTime":"2026-01-22T11:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.545348 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.545608 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.545686 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.545800 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.545878 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:25Z","lastTransitionTime":"2026-01-22T11:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.648269 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.648374 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.648437 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.648472 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.648496 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:25Z","lastTransitionTime":"2026-01-22T11:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.698203 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 04:37:23.510430792 +0000 UTC Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.715577 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:25 crc kubenswrapper[4874]: E0122 11:41:25.716107 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.752055 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.752125 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.752148 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.752178 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.752196 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:25Z","lastTransitionTime":"2026-01-22T11:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.855437 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.855502 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.855519 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.855542 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.855559 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:25Z","lastTransitionTime":"2026-01-22T11:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.958608 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.958653 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.958665 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.958682 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:25 crc kubenswrapper[4874]: I0122 11:41:25.958694 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:25Z","lastTransitionTime":"2026-01-22T11:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.064758 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.065379 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.065543 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.065589 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.065634 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:26Z","lastTransitionTime":"2026-01-22T11:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.168956 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.169001 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.169018 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.169039 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.169055 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:26Z","lastTransitionTime":"2026-01-22T11:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.272251 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.272290 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.272298 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.272312 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.272322 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:26Z","lastTransitionTime":"2026-01-22T11:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.375518 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.375853 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.375998 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.376215 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.376486 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:26Z","lastTransitionTime":"2026-01-22T11:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.479603 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.479635 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.479642 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.479654 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.479662 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:26Z","lastTransitionTime":"2026-01-22T11:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.582041 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.582089 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.582105 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.582130 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.582147 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:26Z","lastTransitionTime":"2026-01-22T11:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.684448 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.684495 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.684507 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.684523 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.684845 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:26Z","lastTransitionTime":"2026-01-22T11:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.699134 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 06:16:43.750801151 +0000 UTC Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.715900 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.716053 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:26 crc kubenswrapper[4874]: E0122 11:41:26.716054 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:26 crc kubenswrapper[4874]: E0122 11:41:26.716126 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.716189 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:26 crc kubenswrapper[4874]: E0122 11:41:26.716241 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.732263 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.757150 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.775598 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.787290 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.787574 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.787775 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.787946 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.788110 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:26Z","lastTransitionTime":"2026-01-22T11:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.792034 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.805835 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.821200 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.835312 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.857057 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:11Z\\\",\\\"message\\\":\\\"actory.go:160\\\\nI0122 11:41:11.510239 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510274 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510315 6499 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 11:41:11.516659 6499 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:41:11.516689 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:41:11.516719 6499 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:41:11.516760 6499 factory.go:656] Stopping watch factory\\\\nI0122 11:41:11.516776 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:41:11.529962 6499 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0122 11:41:11.529996 6499 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0122 11:41:11.530061 6499 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:41:11.530100 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 11:41:11.530194 6499 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.871613 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d542969f-3655-4a4e-8a4d-238cff44f86e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb2908cd58fbcf7ae2f0e4281b7c0ef1a68896ab514d9aa90f347f7346b479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9d6d3c847805c81649e7524a51e0d1a261d3c75d10c90e9e2d6d6a0723ff76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0b6dcbe2e50cd1aa3a3e2edbc5401888ebd6f99cbbf5329245bd4f61bf7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.885789 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.891935 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.892138 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.892268 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.892450 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.892639 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:26Z","lastTransitionTime":"2026-01-22T11:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.898588 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.912738 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.924219 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.934706 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.946839 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.959105 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.971461 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.982548 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:26Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.995702 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.995739 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.995748 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.995762 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:26 crc kubenswrapper[4874]: I0122 11:41:26.995772 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:26Z","lastTransitionTime":"2026-01-22T11:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.097344 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.097375 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.097384 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.097420 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.097435 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:27Z","lastTransitionTime":"2026-01-22T11:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.199084 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.199489 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.199654 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.199777 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.199887 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:27Z","lastTransitionTime":"2026-01-22T11:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.302928 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.302971 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.302984 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.303001 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.303013 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:27Z","lastTransitionTime":"2026-01-22T11:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.405387 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.405435 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.405445 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.405459 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.405470 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:27Z","lastTransitionTime":"2026-01-22T11:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.507810 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.507849 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.507858 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.507872 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.507882 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:27Z","lastTransitionTime":"2026-01-22T11:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.610838 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.610878 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.610889 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.610906 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.610917 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:27Z","lastTransitionTime":"2026-01-22T11:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.699721 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 13:11:26.445714087 +0000 UTC Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.713669 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.713720 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.713732 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.713751 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.713764 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:27Z","lastTransitionTime":"2026-01-22T11:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.715600 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:27 crc kubenswrapper[4874]: E0122 11:41:27.715778 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.816765 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.816801 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.816813 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.816829 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.816841 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:27Z","lastTransitionTime":"2026-01-22T11:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.921021 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.921090 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.921111 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.921133 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:27 crc kubenswrapper[4874]: I0122 11:41:27.921149 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:27Z","lastTransitionTime":"2026-01-22T11:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.030452 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.030522 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.030545 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.030577 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.030600 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.133355 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.133462 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.133487 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.133519 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.133543 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.237828 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.237869 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.237885 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.237900 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.237978 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.342060 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.342123 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.342139 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.342164 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.342182 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.445066 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.445122 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.445142 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.445165 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.445181 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.550025 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.550074 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.550087 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.550108 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.550122 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.653544 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.653604 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.653622 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.653645 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.653661 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.701645 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 18:49:33.73075206 +0000 UTC Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.715380 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.715490 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.715492 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:28 crc kubenswrapper[4874]: E0122 11:41:28.715682 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:28 crc kubenswrapper[4874]: E0122 11:41:28.715821 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:28 crc kubenswrapper[4874]: E0122 11:41:28.716324 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.716953 4874 scope.go:117] "RemoveContainer" containerID="fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337" Jan 22 11:41:28 crc kubenswrapper[4874]: E0122 11:41:28.717381 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.756533 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.756600 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.756627 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.756655 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.756676 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.804270 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.804343 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.804361 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.804383 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.804433 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: E0122 11:41:28.825241 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:28Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.830717 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.830766 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.830779 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.830798 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.830810 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: E0122 11:41:28.848437 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:28Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.853488 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.853564 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.853589 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.853622 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.853644 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: E0122 11:41:28.873850 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:28Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.878953 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.879019 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.879059 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.879094 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.879117 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: E0122 11:41:28.900451 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:28Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.905664 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.905727 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.905749 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.905778 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.905798 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:28 crc kubenswrapper[4874]: E0122 11:41:28.928093 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:28Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:28 crc kubenswrapper[4874]: E0122 11:41:28.928315 4874 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.930470 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.930546 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.930565 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.930994 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:28 crc kubenswrapper[4874]: I0122 11:41:28.931055 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:28Z","lastTransitionTime":"2026-01-22T11:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.033969 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.034008 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.034021 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.034040 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.034054 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:29Z","lastTransitionTime":"2026-01-22T11:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.137466 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.137503 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.137518 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.137536 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.137548 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:29Z","lastTransitionTime":"2026-01-22T11:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.240441 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.240474 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.240487 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.240500 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.240509 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:29Z","lastTransitionTime":"2026-01-22T11:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.343525 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.343564 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.343576 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.343592 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.343603 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:29Z","lastTransitionTime":"2026-01-22T11:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.445948 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.445996 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.446009 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.446028 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.446041 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:29Z","lastTransitionTime":"2026-01-22T11:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.549320 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.549373 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.549385 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.549434 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.549449 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:29Z","lastTransitionTime":"2026-01-22T11:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.652261 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.652343 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.652364 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.652389 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.652439 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:29Z","lastTransitionTime":"2026-01-22T11:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.752692 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 18:06:57.06985107 +0000 UTC Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.753003 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:29 crc kubenswrapper[4874]: E0122 11:41:29.753542 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.754716 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.754764 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.754787 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.754816 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.754839 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:29Z","lastTransitionTime":"2026-01-22T11:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.857802 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.857864 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.857882 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.857907 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.857924 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:29Z","lastTransitionTime":"2026-01-22T11:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.961530 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.961592 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.961603 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.961617 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:29 crc kubenswrapper[4874]: I0122 11:41:29.961627 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:29Z","lastTransitionTime":"2026-01-22T11:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.065145 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.065177 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.065185 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.065201 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.065210 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:30Z","lastTransitionTime":"2026-01-22T11:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.168069 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.168110 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.168119 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.168133 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.168142 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:30Z","lastTransitionTime":"2026-01-22T11:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.270441 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.270475 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.270483 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.270495 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.270504 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:30Z","lastTransitionTime":"2026-01-22T11:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.372827 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.372869 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.372878 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.372894 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.372904 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:30Z","lastTransitionTime":"2026-01-22T11:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.462326 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:30 crc kubenswrapper[4874]: E0122 11:41:30.462657 4874 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:41:30 crc kubenswrapper[4874]: E0122 11:41:30.462820 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs podName:5451fbab-ebad-42e7-bb80-f94bad10d571 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:02.462785416 +0000 UTC m=+96.307856516 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs") pod "network-metrics-daemon-lr2vd" (UID: "5451fbab-ebad-42e7-bb80-f94bad10d571") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.475958 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.476017 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.476034 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.476060 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.476077 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:30Z","lastTransitionTime":"2026-01-22T11:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.578777 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.578819 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.578830 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.578847 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.578856 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:30Z","lastTransitionTime":"2026-01-22T11:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.681133 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.681166 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.681175 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.681189 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.681199 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:30Z","lastTransitionTime":"2026-01-22T11:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.716017 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.716027 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:30 crc kubenswrapper[4874]: E0122 11:41:30.716165 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.716192 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:30 crc kubenswrapper[4874]: E0122 11:41:30.716280 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:30 crc kubenswrapper[4874]: E0122 11:41:30.716412 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.753076 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 07:21:12.169443812 +0000 UTC Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.783384 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.783468 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.783486 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.783509 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.783527 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:30Z","lastTransitionTime":"2026-01-22T11:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.886534 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.886579 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.886592 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.886609 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.886621 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:30Z","lastTransitionTime":"2026-01-22T11:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.989641 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.989680 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.989689 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.989703 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:30 crc kubenswrapper[4874]: I0122 11:41:30.989712 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:30Z","lastTransitionTime":"2026-01-22T11:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.092293 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.092336 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.092364 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.092383 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.092418 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:31Z","lastTransitionTime":"2026-01-22T11:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.194476 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.194526 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.194535 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.194551 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.194561 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:31Z","lastTransitionTime":"2026-01-22T11:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.296863 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.296931 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.296944 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.296994 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.297009 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:31Z","lastTransitionTime":"2026-01-22T11:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.399176 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.399246 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.399259 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.399277 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.399287 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:31Z","lastTransitionTime":"2026-01-22T11:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.501304 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.501341 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.501350 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.501365 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.501374 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:31Z","lastTransitionTime":"2026-01-22T11:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.604411 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.604453 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.604463 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.604478 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.604489 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:31Z","lastTransitionTime":"2026-01-22T11:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.707469 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.707501 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.707509 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.707522 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.707533 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:31Z","lastTransitionTime":"2026-01-22T11:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.716032 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:31 crc kubenswrapper[4874]: E0122 11:41:31.716211 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.754207 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 14:51:58.545358631 +0000 UTC Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.810450 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.810496 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.810511 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.810528 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.810541 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:31Z","lastTransitionTime":"2026-01-22T11:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.912895 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.912943 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.912954 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.912970 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:31 crc kubenswrapper[4874]: I0122 11:41:31.912980 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:31Z","lastTransitionTime":"2026-01-22T11:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.016328 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.016444 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.016466 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.016495 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.016535 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:32Z","lastTransitionTime":"2026-01-22T11:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.098030 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/0.log" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.098079 4874 generic.go:334] "Generic (PLEG): container finished" podID="977746b5-ac1b-4b6e-bdbc-ddd90225e68c" containerID="600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf" exitCode=1 Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.098120 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-krrtc" event={"ID":"977746b5-ac1b-4b6e-bdbc-ddd90225e68c","Type":"ContainerDied","Data":"600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf"} Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.098733 4874 scope.go:117] "RemoveContainer" containerID="600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.112887 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.120586 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.120631 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.120648 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.120671 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.120690 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:32Z","lastTransitionTime":"2026-01-22T11:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.135807 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:11Z\\\",\\\"message\\\":\\\"actory.go:160\\\\nI0122 11:41:11.510239 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510274 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510315 6499 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 11:41:11.516659 6499 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:41:11.516689 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:41:11.516719 6499 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:41:11.516760 6499 factory.go:656] Stopping watch factory\\\\nI0122 11:41:11.516776 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:41:11.529962 6499 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0122 11:41:11.529996 6499 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0122 11:41:11.530061 6499 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:41:11.530100 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 11:41:11.530194 6499 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.154227 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.170583 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d542969f-3655-4a4e-8a4d-238cff44f86e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb2908cd58fbcf7ae2f0e4281b7c0ef1a68896ab514d9aa90f347f7346b479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9d6d3c847805c81649e7524a51e0d1a261d3c75d10c90e9e2d6d6a0723ff76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0b6dcbe2e50cd1aa3a3e2edbc5401888ebd6f99cbbf5329245bd4f61bf7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.184824 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.197059 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:31Z\\\",\\\"message\\\":\\\"2026-01-22T11:40:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90\\\\n2026-01-22T11:40:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90 to /host/opt/cni/bin/\\\\n2026-01-22T11:40:46Z [verbose] multus-daemon started\\\\n2026-01-22T11:40:46Z [verbose] Readiness Indicator file check\\\\n2026-01-22T11:41:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.215238 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.223419 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.223465 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.223478 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.223498 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.223513 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:32Z","lastTransitionTime":"2026-01-22T11:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.228843 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.243656 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.257507 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.268454 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.279731 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.293853 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.324782 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.329149 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.329176 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.329184 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.329197 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.329206 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:32Z","lastTransitionTime":"2026-01-22T11:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.340009 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.351666 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.361043 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.374853 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:32Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.431236 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.431277 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.431286 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.431299 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.431309 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:32Z","lastTransitionTime":"2026-01-22T11:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.533082 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.533110 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.533119 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.533130 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.533139 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:32Z","lastTransitionTime":"2026-01-22T11:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.635022 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.635062 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.635071 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.635086 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.635098 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:32Z","lastTransitionTime":"2026-01-22T11:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.715287 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.715328 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:32 crc kubenswrapper[4874]: E0122 11:41:32.715445 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.715475 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:32 crc kubenswrapper[4874]: E0122 11:41:32.715575 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:32 crc kubenswrapper[4874]: E0122 11:41:32.715644 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.736651 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.736688 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.736699 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.736714 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.736725 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:32Z","lastTransitionTime":"2026-01-22T11:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.755021 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 14:10:53.973580311 +0000 UTC Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.839240 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.839282 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.839294 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.839310 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.839322 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:32Z","lastTransitionTime":"2026-01-22T11:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.942035 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.942066 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.942075 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.942089 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:32 crc kubenswrapper[4874]: I0122 11:41:32.942099 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:32Z","lastTransitionTime":"2026-01-22T11:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.044664 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.044710 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.044722 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.044737 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.044748 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:33Z","lastTransitionTime":"2026-01-22T11:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.104032 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/0.log" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.104087 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-krrtc" event={"ID":"977746b5-ac1b-4b6e-bdbc-ddd90225e68c","Type":"ContainerStarted","Data":"cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8"} Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.118876 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.131604 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.147551 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.147591 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.147601 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.147617 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.147628 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:33Z","lastTransitionTime":"2026-01-22T11:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.148612 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.167603 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.181324 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.197632 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.208076 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.225783 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.251973 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.252015 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.252024 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.252042 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.252059 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:33Z","lastTransitionTime":"2026-01-22T11:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.256898 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.282376 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:11Z\\\",\\\"message\\\":\\\"actory.go:160\\\\nI0122 11:41:11.510239 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510274 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510315 6499 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 11:41:11.516659 6499 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:41:11.516689 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:41:11.516719 6499 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:41:11.516760 6499 factory.go:656] Stopping watch factory\\\\nI0122 11:41:11.516776 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:41:11.529962 6499 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0122 11:41:11.529996 6499 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0122 11:41:11.530061 6499 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:41:11.530100 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 11:41:11.530194 6499 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.297137 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.309836 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.321940 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.333518 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:31Z\\\",\\\"message\\\":\\\"2026-01-22T11:40:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90\\\\n2026-01-22T11:40:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90 to /host/opt/cni/bin/\\\\n2026-01-22T11:40:46Z [verbose] multus-daemon started\\\\n2026-01-22T11:40:46Z [verbose] Readiness Indicator file check\\\\n2026-01-22T11:41:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.345780 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.354484 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.354516 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.354527 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.354541 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.354564 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:33Z","lastTransitionTime":"2026-01-22T11:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.356793 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.369161 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.381154 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d542969f-3655-4a4e-8a4d-238cff44f86e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb2908cd58fbcf7ae2f0e4281b7c0ef1a68896ab514d9aa90f347f7346b479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9d6d3c847805c81649e7524a51e0d1a261d3c75d10c90e9e2d6d6a0723ff76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0b6dcbe2e50cd1aa3a3e2edbc5401888ebd6f99cbbf5329245bd4f61bf7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:33Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.456870 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.456939 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.456961 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.456990 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.457010 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:33Z","lastTransitionTime":"2026-01-22T11:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.559196 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.559235 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.559248 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.559264 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.559275 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:33Z","lastTransitionTime":"2026-01-22T11:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.661590 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.661628 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.661639 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.661654 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.661664 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:33Z","lastTransitionTime":"2026-01-22T11:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.715575 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:33 crc kubenswrapper[4874]: E0122 11:41:33.715799 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.755498 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 07:46:59.084904856 +0000 UTC Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.764503 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.764561 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.764605 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.764624 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.764635 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:33Z","lastTransitionTime":"2026-01-22T11:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.866838 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.866870 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.866881 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.866898 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.866910 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:33Z","lastTransitionTime":"2026-01-22T11:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.969347 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.969385 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.969410 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.969425 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:33 crc kubenswrapper[4874]: I0122 11:41:33.969436 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:33Z","lastTransitionTime":"2026-01-22T11:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.071646 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.071699 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.071721 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.071749 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.071775 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:34Z","lastTransitionTime":"2026-01-22T11:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.174215 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.174292 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.174319 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.174350 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.174373 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:34Z","lastTransitionTime":"2026-01-22T11:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.277229 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.277308 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.277329 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.277353 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.277375 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:34Z","lastTransitionTime":"2026-01-22T11:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.380281 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.380329 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.380340 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.380360 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.380369 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:34Z","lastTransitionTime":"2026-01-22T11:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.483039 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.483074 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.483082 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.483096 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.483105 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:34Z","lastTransitionTime":"2026-01-22T11:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.585692 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.585730 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.585739 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.585752 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.585760 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:34Z","lastTransitionTime":"2026-01-22T11:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.688621 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.688715 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.688737 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.688772 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.688796 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:34Z","lastTransitionTime":"2026-01-22T11:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.715435 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.715540 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:34 crc kubenswrapper[4874]: E0122 11:41:34.715614 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.715728 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:34 crc kubenswrapper[4874]: E0122 11:41:34.715830 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:34 crc kubenswrapper[4874]: E0122 11:41:34.715965 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.755763 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 05:29:39.66697323 +0000 UTC Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.791262 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.791315 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.791331 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.791353 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.791369 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:34Z","lastTransitionTime":"2026-01-22T11:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.893844 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.893896 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.893910 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.893950 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.893968 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:34Z","lastTransitionTime":"2026-01-22T11:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.996148 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.996195 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.996209 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.996225 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:34 crc kubenswrapper[4874]: I0122 11:41:34.996237 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:34Z","lastTransitionTime":"2026-01-22T11:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.098303 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.098351 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.098366 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.098388 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.098428 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:35Z","lastTransitionTime":"2026-01-22T11:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.200915 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.200970 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.200986 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.201006 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.201024 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:35Z","lastTransitionTime":"2026-01-22T11:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.302872 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.302919 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.302932 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.302951 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.302963 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:35Z","lastTransitionTime":"2026-01-22T11:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.405428 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.405485 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.405497 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.405512 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.405521 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:35Z","lastTransitionTime":"2026-01-22T11:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.507352 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.507408 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.507425 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.507440 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.507449 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:35Z","lastTransitionTime":"2026-01-22T11:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.608963 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.609005 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.609019 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.609034 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.609045 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:35Z","lastTransitionTime":"2026-01-22T11:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.710997 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.711048 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.711064 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.711088 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.711103 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:35Z","lastTransitionTime":"2026-01-22T11:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.715293 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:35 crc kubenswrapper[4874]: E0122 11:41:35.715645 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.756133 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 10:47:21.635014387 +0000 UTC Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.813705 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.813750 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.813762 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.813779 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.813791 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:35Z","lastTransitionTime":"2026-01-22T11:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.915682 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.915728 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.915740 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.915757 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:35 crc kubenswrapper[4874]: I0122 11:41:35.915767 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:35Z","lastTransitionTime":"2026-01-22T11:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.017848 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.017876 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.017887 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.017901 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.017912 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:36Z","lastTransitionTime":"2026-01-22T11:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.120347 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.120416 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.120435 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.120457 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.120472 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:36Z","lastTransitionTime":"2026-01-22T11:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.223209 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.223243 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.223252 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.223269 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.223278 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:36Z","lastTransitionTime":"2026-01-22T11:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.325747 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.325790 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.325801 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.325819 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.325833 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:36Z","lastTransitionTime":"2026-01-22T11:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.427978 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.428038 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.428048 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.428062 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.428071 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:36Z","lastTransitionTime":"2026-01-22T11:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.531028 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.531064 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.531073 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.531087 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.531096 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:36Z","lastTransitionTime":"2026-01-22T11:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.632975 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.633017 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.633028 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.633043 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.633055 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:36Z","lastTransitionTime":"2026-01-22T11:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.715686 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.715727 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.715743 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:36 crc kubenswrapper[4874]: E0122 11:41:36.716109 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:36 crc kubenswrapper[4874]: E0122 11:41:36.716206 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:36 crc kubenswrapper[4874]: E0122 11:41:36.715968 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.726496 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.737904 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.737946 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.737956 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.737977 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.737990 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:36Z","lastTransitionTime":"2026-01-22T11:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.747379 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.756757 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 02:38:42.088652766 +0000 UTC Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.762800 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.775650 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.784786 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.796628 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.809231 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.830931 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:11Z\\\",\\\"message\\\":\\\"actory.go:160\\\\nI0122 11:41:11.510239 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510274 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510315 6499 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 11:41:11.516659 6499 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:41:11.516689 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:41:11.516719 6499 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:41:11.516760 6499 factory.go:656] Stopping watch factory\\\\nI0122 11:41:11.516776 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:41:11.529962 6499 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0122 11:41:11.529996 6499 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0122 11:41:11.530061 6499 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:41:11.530100 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 11:41:11.530194 6499 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.839762 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.839805 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.839816 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.839835 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.839847 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:36Z","lastTransitionTime":"2026-01-22T11:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.842057 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d542969f-3655-4a4e-8a4d-238cff44f86e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb2908cd58fbcf7ae2f0e4281b7c0ef1a68896ab514d9aa90f347f7346b479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9d6d3c847805c81649e7524a51e0d1a261d3c75d10c90e9e2d6d6a0723ff76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0b6dcbe2e50cd1aa3a3e2edbc5401888ebd6f99cbbf5329245bd4f61bf7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.853790 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.865639 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:31Z\\\",\\\"message\\\":\\\"2026-01-22T11:40:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90\\\\n2026-01-22T11:40:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90 to /host/opt/cni/bin/\\\\n2026-01-22T11:40:46Z [verbose] multus-daemon started\\\\n2026-01-22T11:40:46Z [verbose] Readiness Indicator file check\\\\n2026-01-22T11:41:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.878445 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.888487 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.899030 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.909832 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.921324 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.932355 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.941847 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.941886 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.941898 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.941915 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.941926 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:36Z","lastTransitionTime":"2026-01-22T11:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:36 crc kubenswrapper[4874]: I0122 11:41:36.942727 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:36Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.043597 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.043627 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.043636 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.043650 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.043662 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:37Z","lastTransitionTime":"2026-01-22T11:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.145936 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.146011 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.146022 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.146036 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.146045 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:37Z","lastTransitionTime":"2026-01-22T11:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.248353 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.248415 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.248427 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.248441 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.248453 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:37Z","lastTransitionTime":"2026-01-22T11:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.365992 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.366025 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.366033 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.366046 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.366054 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:37Z","lastTransitionTime":"2026-01-22T11:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.468489 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.468545 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.468560 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.468582 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.468601 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:37Z","lastTransitionTime":"2026-01-22T11:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.571641 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.571696 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.571708 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.571727 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.571742 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:37Z","lastTransitionTime":"2026-01-22T11:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.675125 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.675169 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.675180 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.675195 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.675209 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:37Z","lastTransitionTime":"2026-01-22T11:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.715533 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:37 crc kubenswrapper[4874]: E0122 11:41:37.715642 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.757916 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 06:19:36.955140537 +0000 UTC Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.777633 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.777701 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.777723 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.777754 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.777777 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:37Z","lastTransitionTime":"2026-01-22T11:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.879915 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.880004 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.880027 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.880053 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.880072 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:37Z","lastTransitionTime":"2026-01-22T11:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.981737 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.981767 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.981776 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.981788 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:37 crc kubenswrapper[4874]: I0122 11:41:37.981796 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:37Z","lastTransitionTime":"2026-01-22T11:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.084197 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.084224 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.084232 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.084244 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.084253 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:38Z","lastTransitionTime":"2026-01-22T11:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.186184 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.186222 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.186233 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.186249 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.186260 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:38Z","lastTransitionTime":"2026-01-22T11:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.288192 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.288228 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.288239 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.288256 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.288268 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:38Z","lastTransitionTime":"2026-01-22T11:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.391303 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.391344 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.391352 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.391366 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.391375 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:38Z","lastTransitionTime":"2026-01-22T11:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.493508 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.493741 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.493843 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.493928 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.493996 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:38Z","lastTransitionTime":"2026-01-22T11:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.596511 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.596889 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.597030 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.597139 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.597227 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:38Z","lastTransitionTime":"2026-01-22T11:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.699863 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.699931 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.699943 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.699959 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.699969 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:38Z","lastTransitionTime":"2026-01-22T11:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.715494 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.715543 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:38 crc kubenswrapper[4874]: E0122 11:41:38.715601 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:38 crc kubenswrapper[4874]: E0122 11:41:38.715691 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.715778 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:38 crc kubenswrapper[4874]: E0122 11:41:38.715836 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.758833 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 13:15:36.769117105 +0000 UTC Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.803114 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.803153 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.803161 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.803177 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.803191 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:38Z","lastTransitionTime":"2026-01-22T11:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.905671 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.905701 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.905711 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.905743 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:38 crc kubenswrapper[4874]: I0122 11:41:38.905753 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:38Z","lastTransitionTime":"2026-01-22T11:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.008307 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.008420 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.008609 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.008636 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.008649 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.057455 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.057487 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.057496 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.057509 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.057520 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: E0122 11:41:39.070641 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:39Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.074326 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.074367 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.074378 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.074408 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.074422 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: E0122 11:41:39.089688 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:39Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.092970 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.093014 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.093030 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.093051 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.093066 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: E0122 11:41:39.106764 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:39Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.109737 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.109829 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.109908 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.109980 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.110067 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: E0122 11:41:39.123156 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:39Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.126063 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.126092 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.126101 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.126114 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.126138 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: E0122 11:41:39.138849 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:39Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:39 crc kubenswrapper[4874]: E0122 11:41:39.138993 4874 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.140231 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.140261 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.140271 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.140286 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.140296 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.242639 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.242666 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.242675 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.242689 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.242699 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.344306 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.344331 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.344338 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.344351 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.344360 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.446381 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.446426 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.446433 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.446449 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.446459 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.548906 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.548944 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.548954 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.548972 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.548983 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.651670 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.651716 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.651727 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.651741 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.651752 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.715266 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:39 crc kubenswrapper[4874]: E0122 11:41:39.715509 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.716085 4874 scope.go:117] "RemoveContainer" containerID="fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.727319 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.753193 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.753229 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.753238 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.753252 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.753262 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.759804 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 19:17:52.16725777 +0000 UTC Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.855515 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.855842 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.855851 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.855864 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.855874 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.958567 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.958621 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.958630 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.958647 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:39 crc kubenswrapper[4874]: I0122 11:41:39.958656 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:39Z","lastTransitionTime":"2026-01-22T11:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.060650 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.060681 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.060689 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.060701 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.060710 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:40Z","lastTransitionTime":"2026-01-22T11:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.127210 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/2.log" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.136719 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35"} Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.137292 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.152727 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.163236 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.163287 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.163304 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.163325 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.163338 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:40Z","lastTransitionTime":"2026-01-22T11:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.175967 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.187325 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.198489 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.207746 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.220894 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.236134 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.260755 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:11Z\\\",\\\"message\\\":\\\"actory.go:160\\\\nI0122 11:41:11.510239 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510274 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510315 6499 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 11:41:11.516659 6499 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:41:11.516689 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:41:11.516719 6499 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:41:11.516760 6499 factory.go:656] Stopping watch factory\\\\nI0122 11:41:11.516776 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:41:11.529962 6499 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0122 11:41:11.529996 6499 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0122 11:41:11.530061 6499 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:41:11.530100 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 11:41:11.530194 6499 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:41:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.265331 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.265367 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.265380 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.265411 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.265424 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:40Z","lastTransitionTime":"2026-01-22T11:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.274021 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb774505-642d-49da-a7c9-20fc8991d5bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ca3abbae7b47aa2ee502ed6cc36a325843a9f44e1c7881ba5a142bd13dd1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b54087988f1e51ce17beb3055e35b3ff31a6aa3cc3d687a18a7f6afdf9505e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b54087988f1e51ce17beb3055e35b3ff31a6aa3cc3d687a18a7f6afdf9505e4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.294710 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d542969f-3655-4a4e-8a4d-238cff44f86e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb2908cd58fbcf7ae2f0e4281b7c0ef1a68896ab514d9aa90f347f7346b479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9d6d3c847805c81649e7524a51e0d1a261d3c75d10c90e9e2d6d6a0723ff76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0b6dcbe2e50cd1aa3a3e2edbc5401888ebd6f99cbbf5329245bd4f61bf7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.311000 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.322642 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:31Z\\\",\\\"message\\\":\\\"2026-01-22T11:40:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90\\\\n2026-01-22T11:40:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90 to /host/opt/cni/bin/\\\\n2026-01-22T11:40:46Z [verbose] multus-daemon started\\\\n2026-01-22T11:40:46Z [verbose] Readiness Indicator file check\\\\n2026-01-22T11:41:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.337839 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.347314 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.356667 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.367165 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.367200 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.367211 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.367227 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.367240 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:40Z","lastTransitionTime":"2026-01-22T11:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.367450 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.378180 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.387481 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.396209 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:40Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.469653 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.469689 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.469702 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.469718 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.469731 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:40Z","lastTransitionTime":"2026-01-22T11:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.572439 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.572484 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.572506 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.572520 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.572528 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:40Z","lastTransitionTime":"2026-01-22T11:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.674268 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.674308 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.674321 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.674336 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.674348 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:40Z","lastTransitionTime":"2026-01-22T11:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.715909 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.715955 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.715909 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:40 crc kubenswrapper[4874]: E0122 11:41:40.716046 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:40 crc kubenswrapper[4874]: E0122 11:41:40.716100 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:40 crc kubenswrapper[4874]: E0122 11:41:40.716142 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.759900 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 10:03:09.524152126 +0000 UTC Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.777050 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.777075 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.777083 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.777095 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.777103 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:40Z","lastTransitionTime":"2026-01-22T11:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.879341 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.879373 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.879382 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.879424 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.879433 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:40Z","lastTransitionTime":"2026-01-22T11:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.981593 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.981656 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.981668 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.981685 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:40 crc kubenswrapper[4874]: I0122 11:41:40.981697 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:40Z","lastTransitionTime":"2026-01-22T11:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.083883 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.083911 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.083920 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.083932 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.083940 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:41Z","lastTransitionTime":"2026-01-22T11:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.140841 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/3.log" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.141253 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/2.log" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.143071 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" exitCode=1 Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.143103 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35"} Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.143132 4874 scope.go:117] "RemoveContainer" containerID="fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.143731 4874 scope.go:117] "RemoveContainer" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:41:41 crc kubenswrapper[4874]: E0122 11:41:41.143858 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.159698 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.175236 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.186452 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.191682 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.191708 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.191718 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.191732 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.191741 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:41Z","lastTransitionTime":"2026-01-22T11:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.203158 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.216980 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.242526 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.254691 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.265879 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.279730 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.294479 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.294554 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.294568 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.294880 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.294895 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:41Z","lastTransitionTime":"2026-01-22T11:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.296175 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.310222 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.327557 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa3afc6324d7be4bb8f25f93db31da4205d1cf4c8956bbee8fc391d2fe99d337\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:11Z\\\",\\\"message\\\":\\\"actory.go:160\\\\nI0122 11:41:11.510239 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510274 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 11:41:11.510315 6499 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 11:41:11.516659 6499 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 11:41:11.516689 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 11:41:11.516719 6499 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 11:41:11.516760 6499 factory.go:656] Stopping watch factory\\\\nI0122 11:41:11.516776 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 11:41:11.529962 6499 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0122 11:41:11.529996 6499 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0122 11:41:11.530061 6499 ovnkube.go:599] Stopped ovnkube\\\\nI0122 11:41:11.530100 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 11:41:11.530194 6499 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:40Z\\\",\\\"message\\\":\\\"GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 11:41:40.694245 6904 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 11:41:40.694231 6904 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.339224 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.348044 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb774505-642d-49da-a7c9-20fc8991d5bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ca3abbae7b47aa2ee502ed6cc36a325843a9f44e1c7881ba5a142bd13dd1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b54087988f1e51ce17beb3055e35b3ff31a6aa3cc3d687a18a7f6afdf9505e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b54087988f1e51ce17beb3055e35b3ff31a6aa3cc3d687a18a7f6afdf9505e4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.357777 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d542969f-3655-4a4e-8a4d-238cff44f86e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb2908cd58fbcf7ae2f0e4281b7c0ef1a68896ab514d9aa90f347f7346b479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9d6d3c847805c81649e7524a51e0d1a261d3c75d10c90e9e2d6d6a0723ff76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0b6dcbe2e50cd1aa3a3e2edbc5401888ebd6f99cbbf5329245bd4f61bf7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.368004 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.377958 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:31Z\\\",\\\"message\\\":\\\"2026-01-22T11:40:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90\\\\n2026-01-22T11:40:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90 to /host/opt/cni/bin/\\\\n2026-01-22T11:40:46Z [verbose] multus-daemon started\\\\n2026-01-22T11:40:46Z [verbose] Readiness Indicator file check\\\\n2026-01-22T11:41:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.390427 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.396718 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.396752 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.396762 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.396775 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.396783 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:41Z","lastTransitionTime":"2026-01-22T11:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.399729 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:41Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.499829 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.499864 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.499873 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.499887 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.499903 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:41Z","lastTransitionTime":"2026-01-22T11:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.602083 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.602109 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.602118 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.602133 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.602143 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:41Z","lastTransitionTime":"2026-01-22T11:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.704779 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.704858 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.704882 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.704911 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.704933 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:41Z","lastTransitionTime":"2026-01-22T11:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.715344 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:41 crc kubenswrapper[4874]: E0122 11:41:41.715560 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.760470 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 01:52:05.533615663 +0000 UTC Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.807984 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.808020 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.808033 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.808049 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.808062 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:41Z","lastTransitionTime":"2026-01-22T11:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.910951 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.910996 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.911010 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.911031 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:41 crc kubenswrapper[4874]: I0122 11:41:41.911046 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:41Z","lastTransitionTime":"2026-01-22T11:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.013539 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.013572 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.013581 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.013595 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.013604 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:42Z","lastTransitionTime":"2026-01-22T11:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.116220 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.116260 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.116269 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.116286 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.116296 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:42Z","lastTransitionTime":"2026-01-22T11:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.148109 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/3.log" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.152998 4874 scope.go:117] "RemoveContainer" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:41:42 crc kubenswrapper[4874]: E0122 11:41:42.153199 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.169149 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.183649 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.212889 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:40Z\\\",\\\"message\\\":\\\"GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 11:41:40.694245 6904 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 11:41:40.694231 6904 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.218437 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.218476 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.218491 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.218509 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.218523 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:42Z","lastTransitionTime":"2026-01-22T11:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.223770 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb774505-642d-49da-a7c9-20fc8991d5bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ca3abbae7b47aa2ee502ed6cc36a325843a9f44e1c7881ba5a142bd13dd1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b54087988f1e51ce17beb3055e35b3ff31a6aa3cc3d687a18a7f6afdf9505e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b54087988f1e51ce17beb3055e35b3ff31a6aa3cc3d687a18a7f6afdf9505e4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.234927 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d542969f-3655-4a4e-8a4d-238cff44f86e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb2908cd58fbcf7ae2f0e4281b7c0ef1a68896ab514d9aa90f347f7346b479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9d6d3c847805c81649e7524a51e0d1a261d3c75d10c90e9e2d6d6a0723ff76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0b6dcbe2e50cd1aa3a3e2edbc5401888ebd6f99cbbf5329245bd4f61bf7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.245035 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.261449 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:31Z\\\",\\\"message\\\":\\\"2026-01-22T11:40:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90\\\\n2026-01-22T11:40:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90 to /host/opt/cni/bin/\\\\n2026-01-22T11:40:46Z [verbose] multus-daemon started\\\\n2026-01-22T11:40:46Z [verbose] Readiness Indicator file check\\\\n2026-01-22T11:41:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.276589 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.289382 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.300137 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.311772 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.321090 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.321133 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.321146 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.321163 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.321175 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:42Z","lastTransitionTime":"2026-01-22T11:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.323662 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.334392 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.346910 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.361466 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.378007 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.391123 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.404140 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.449248 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.449289 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.449301 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.449326 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.449339 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:42Z","lastTransitionTime":"2026-01-22T11:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.453805 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:42Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.551852 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.551899 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.551912 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.551928 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.551939 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:42Z","lastTransitionTime":"2026-01-22T11:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.654416 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.654451 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.654460 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.654473 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.654482 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:42Z","lastTransitionTime":"2026-01-22T11:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.715282 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.715318 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.715350 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:42 crc kubenswrapper[4874]: E0122 11:41:42.715450 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:42 crc kubenswrapper[4874]: E0122 11:41:42.715521 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:42 crc kubenswrapper[4874]: E0122 11:41:42.715601 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.757281 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.757317 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.757328 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.757343 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.757354 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:42Z","lastTransitionTime":"2026-01-22T11:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.761067 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 19:17:29.065492057 +0000 UTC Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.860688 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.860729 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.860741 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.860757 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.860768 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:42Z","lastTransitionTime":"2026-01-22T11:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.962831 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.962881 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.962892 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.962909 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:42 crc kubenswrapper[4874]: I0122 11:41:42.962921 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:42Z","lastTransitionTime":"2026-01-22T11:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.065729 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.065787 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.065809 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.065837 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.065858 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:43Z","lastTransitionTime":"2026-01-22T11:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.168583 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.168642 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.168660 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.168689 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.168713 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:43Z","lastTransitionTime":"2026-01-22T11:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.271604 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.271661 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.271681 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.271707 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.271724 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:43Z","lastTransitionTime":"2026-01-22T11:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.374816 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.374850 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.374871 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.374897 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.374914 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:43Z","lastTransitionTime":"2026-01-22T11:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.478172 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.478633 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.478801 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.478978 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.479115 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:43Z","lastTransitionTime":"2026-01-22T11:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.581230 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.581267 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.581276 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.581291 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.581332 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:43Z","lastTransitionTime":"2026-01-22T11:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.684394 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.684825 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.684994 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.685145 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.685309 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:43Z","lastTransitionTime":"2026-01-22T11:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.716142 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:43 crc kubenswrapper[4874]: E0122 11:41:43.716577 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.761766 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 14:45:36.839439613 +0000 UTC Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.789008 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.789350 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.789584 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.789861 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.790070 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:43Z","lastTransitionTime":"2026-01-22T11:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.893667 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.893736 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.893756 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.893781 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.893801 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:43Z","lastTransitionTime":"2026-01-22T11:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.996949 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.997002 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.997017 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.997043 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:43 crc kubenswrapper[4874]: I0122 11:41:43.997062 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:43Z","lastTransitionTime":"2026-01-22T11:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.099627 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.099691 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.099710 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.099736 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.099755 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:44Z","lastTransitionTime":"2026-01-22T11:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.204013 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.204116 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.204183 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.204223 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.204248 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:44Z","lastTransitionTime":"2026-01-22T11:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.307854 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.307953 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.307973 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.308001 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.308019 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:44Z","lastTransitionTime":"2026-01-22T11:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.410700 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.410765 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.410803 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.410838 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.410862 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:44Z","lastTransitionTime":"2026-01-22T11:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.514131 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.514200 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.514216 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.514242 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.514258 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:44Z","lastTransitionTime":"2026-01-22T11:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.617078 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.617155 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.617177 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.617204 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.617225 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:44Z","lastTransitionTime":"2026-01-22T11:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.715925 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.716059 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.716175 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:44 crc kubenswrapper[4874]: E0122 11:41:44.716190 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:44 crc kubenswrapper[4874]: E0122 11:41:44.716363 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:44 crc kubenswrapper[4874]: E0122 11:41:44.716577 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.719962 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.719996 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.720004 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.720019 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.720029 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:44Z","lastTransitionTime":"2026-01-22T11:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.762834 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 10:49:04.796117869 +0000 UTC Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.822565 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.822655 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.822672 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.822697 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.822717 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:44Z","lastTransitionTime":"2026-01-22T11:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.925706 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.925768 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.925787 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.925817 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:44 crc kubenswrapper[4874]: I0122 11:41:44.925839 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:44Z","lastTransitionTime":"2026-01-22T11:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.028860 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.028920 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.028937 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.028964 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.028980 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:45Z","lastTransitionTime":"2026-01-22T11:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.131574 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.131644 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.131655 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.131671 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.131683 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:45Z","lastTransitionTime":"2026-01-22T11:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.234938 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.234986 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.235002 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.235024 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.235040 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:45Z","lastTransitionTime":"2026-01-22T11:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.344851 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.344948 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.344976 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.345011 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.345035 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:45Z","lastTransitionTime":"2026-01-22T11:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.447731 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.447769 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.447780 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.447797 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.447807 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:45Z","lastTransitionTime":"2026-01-22T11:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.550717 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.550793 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.550803 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.550825 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.550838 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:45Z","lastTransitionTime":"2026-01-22T11:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.654299 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.654341 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.654352 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.654369 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.654571 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:45Z","lastTransitionTime":"2026-01-22T11:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.715072 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:45 crc kubenswrapper[4874]: E0122 11:41:45.715256 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.757366 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.757445 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.757459 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.757476 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.757487 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:45Z","lastTransitionTime":"2026-01-22T11:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.763646 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 06:12:48.991125852 +0000 UTC Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.860560 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.860619 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.860641 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.860665 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.860682 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:45Z","lastTransitionTime":"2026-01-22T11:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.963995 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.964069 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.964094 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.964125 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:45 crc kubenswrapper[4874]: I0122 11:41:45.964147 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:45Z","lastTransitionTime":"2026-01-22T11:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.504350 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.504383 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.504409 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.504424 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.504437 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:46Z","lastTransitionTime":"2026-01-22T11:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.606815 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.606864 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.606875 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.606893 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.606905 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:46Z","lastTransitionTime":"2026-01-22T11:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.710122 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.710178 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.710190 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.710209 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.710224 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:46Z","lastTransitionTime":"2026-01-22T11:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.715736 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:46 crc kubenswrapper[4874]: E0122 11:41:46.715877 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.716014 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.716022 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:46 crc kubenswrapper[4874]: E0122 11:41:46.716213 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:46 crc kubenswrapper[4874]: E0122 11:41:46.716370 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.733494 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.755694 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.763751 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 12:44:40.088664559 +0000 UTC Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.780013 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:40Z\\\",\\\"message\\\":\\\"GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 11:41:40.694245 6904 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 11:41:40.694231 6904 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:41:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9lv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6tmll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.792609 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q2rnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561c457d-4767-4b66-a07a-c435b7c9f161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e3f356aa5539627ad44785049bfc3d7ac69b53acb6b900b6e7f568ea776db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn6zq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q2rnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.808200 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8470cb5-cfaf-4760-8c07-ce375052950f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851de9bc28a46137d2c81498f2c5b5814139e518b3f0d2bf9e78a7928825f7c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0b02c44c270a461f55a331b0c381f4287009b626412019fc8b109a0e9c330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9smw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5vd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.813476 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.813532 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.813552 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.813579 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.813597 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:46Z","lastTransitionTime":"2026-01-22T11:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.827494 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb774505-642d-49da-a7c9-20fc8991d5bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ca3abbae7b47aa2ee502ed6cc36a325843a9f44e1c7881ba5a142bd13dd1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b54087988f1e51ce17beb3055e35b3ff31a6aa3cc3d687a18a7f6afdf9505e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b54087988f1e51ce17beb3055e35b3ff31a6aa3cc3d687a18a7f6afdf9505e4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.843051 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d542969f-3655-4a4e-8a4d-238cff44f86e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9feb2908cd58fbcf7ae2f0e4281b7c0ef1a68896ab514d9aa90f347f7346b479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9d6d3c847805c81649e7524a51e0d1a261d3c75d10c90e9e2d6d6a0723ff76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0b6dcbe2e50cd1aa3a3e2edbc5401888ebd6f99cbbf5329245bd4f61bf7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a089dbfb63c0a243eb60e931dbdef0c66dac617d8da395573030e5cb3c6a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.857755 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.875434 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-krrtc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"977746b5-ac1b-4b6e-bdbc-ddd90225e68c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T11:41:31Z\\\",\\\"message\\\":\\\"2026-01-22T11:40:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90\\\\n2026-01-22T11:40:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4dce4b60-5f3b-4a6c-a314-b1ffb3916b90 to /host/opt/cni/bin/\\\\n2026-01-22T11:40:46Z [verbose] multus-daemon started\\\\n2026-01-22T11:40:46Z [verbose] Readiness Indicator file check\\\\n2026-01-22T11:41:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssgjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-krrtc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.891732 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0e516a3e013fb8d23955f3170640da33bd3342dd094539b0ce10e6282761907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b277eb15265649871e04b7326942a50935322dbf06b2980b50dca2c681950c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d81a42eb7434f8c47532aaa5fc067af6db074f1966ca316b5f906e4a5315819a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a89a876896ac19a81f506739c3169b672dd0f2359d3c11bf37500bd7d7b56f14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c1dea0ccef6db4e333abbfcdaf54e048a6860bb27c8863d4282424ab0ef6f63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdf62f50193de58be29655476074eda616b705e5fe64dae97bfda7aca5a690e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617ad7b2c963ff9975f51646e9d86516c895a9bea66823867fc458e199eb1d94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txncl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pdb2m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.907588 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34389610bb49cff50f1cc1d4eecd26651272b1a0d761d312ab63dc10dcec5176\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.916584 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.916627 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.916638 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.916655 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.916666 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:46Z","lastTransitionTime":"2026-01-22T11:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.921207 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.933774 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-prbck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5ca785e-1db4-4e08-9ad0-66158728b48a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aebd343dfc49cbbea157cbfaed8299a4a619ef02f521f13f073887c989b0295f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7628z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-prbck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.943713 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c9653f9-cd5b-4b7a-8056-80ae8235d039\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac071038417ac8cf2cbb024f65f2d0c0177a34889700e0f690d5013f186236b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4prkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.964943 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:46 crc kubenswrapper[4874]: I0122 11:41:46.997382 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:46Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.011464 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.020260 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.020320 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.020334 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.020357 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.020370 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:47Z","lastTransitionTime":"2026-01-22T11:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.028055 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.043633 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:47Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.122820 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.122868 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.122880 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.122896 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.122907 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:47Z","lastTransitionTime":"2026-01-22T11:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.224767 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.224811 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.224820 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.224833 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.224841 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:47Z","lastTransitionTime":"2026-01-22T11:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.327997 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.328050 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.328067 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.328085 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.328096 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:47Z","lastTransitionTime":"2026-01-22T11:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.431297 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.431364 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.431378 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.431419 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.431435 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:47Z","lastTransitionTime":"2026-01-22T11:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.534661 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.534714 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.534724 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.534744 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.534755 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:47Z","lastTransitionTime":"2026-01-22T11:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.637990 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.638030 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.638040 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.638057 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.638075 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:47Z","lastTransitionTime":"2026-01-22T11:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.715471 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:47 crc kubenswrapper[4874]: E0122 11:41:47.715674 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.740956 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.740998 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.741009 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.741024 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.741037 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:47Z","lastTransitionTime":"2026-01-22T11:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.764476 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 13:29:59.780441503 +0000 UTC Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.843892 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.843943 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.843957 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.843978 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.843999 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:47Z","lastTransitionTime":"2026-01-22T11:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.946788 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.946814 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.946822 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.946836 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:47 crc kubenswrapper[4874]: I0122 11:41:47.946846 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:47Z","lastTransitionTime":"2026-01-22T11:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.050228 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.050297 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.050319 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.050347 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.050366 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:48Z","lastTransitionTime":"2026-01-22T11:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.152980 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.153022 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.153032 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.153050 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.153061 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:48Z","lastTransitionTime":"2026-01-22T11:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.254987 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.255038 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.255051 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.255068 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.255081 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:48Z","lastTransitionTime":"2026-01-22T11:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.357705 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.357763 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.357775 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.357794 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.357812 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:48Z","lastTransitionTime":"2026-01-22T11:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.460180 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.460255 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.460265 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.460284 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.460293 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:48Z","lastTransitionTime":"2026-01-22T11:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.562360 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.562417 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.562431 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.562452 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.562464 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:48Z","lastTransitionTime":"2026-01-22T11:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.665594 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.665625 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.665636 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.665652 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.665664 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:48Z","lastTransitionTime":"2026-01-22T11:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.715926 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.716069 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.716084 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.716087 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.716264 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.716383 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.730524 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.730643 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.730669 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.730686 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.730740 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:52.730714819 +0000 UTC m=+146.575785919 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.730796 4874 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.730851 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.730796 4874 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.730872 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.730885 4874 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.730855 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:52.730837923 +0000 UTC m=+146.575908993 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.730941 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:52.730929776 +0000 UTC m=+146.576000946 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.730965 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:52.730958237 +0000 UTC m=+146.576029307 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.764854 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 07:18:41.8273358 +0000 UTC Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.767559 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.767585 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.767596 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.767612 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.767623 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:48Z","lastTransitionTime":"2026-01-22T11:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.832144 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.832290 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.832308 4874 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.832320 4874 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:48 crc kubenswrapper[4874]: E0122 11:41:48.832372 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:52.832356286 +0000 UTC m=+146.677427356 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.870510 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.870589 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.870614 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.870644 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.870666 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:48Z","lastTransitionTime":"2026-01-22T11:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.973582 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.973643 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.973664 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.973691 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:48 crc kubenswrapper[4874]: I0122 11:41:48.973711 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:48Z","lastTransitionTime":"2026-01-22T11:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.080265 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.080312 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.080323 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.080340 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.080352 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.167608 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.167801 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.167830 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.167904 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.167927 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: E0122 11:41:49.188627 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.193125 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.193177 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.193195 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.193218 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.193234 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: E0122 11:41:49.216511 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.222264 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.222292 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.222303 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.222334 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.222346 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: E0122 11:41:49.240804 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.244937 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.244981 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.244992 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.245010 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.245023 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: E0122 11:41:49.262701 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.267203 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.267227 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.267235 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.267247 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.267257 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: E0122 11:41:49.281873 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T11:41:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"770ef4c0-49b6-4adf-aa62-b643a71c762c\\\",\\\"systemUUID\\\":\\\"400a411a-8387-4bfb-bbce-2d30a7ad1d2e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:49Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:49 crc kubenswrapper[4874]: E0122 11:41:49.282052 4874 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.283906 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.283946 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.283961 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.283977 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.283990 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.387263 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.387321 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.387337 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.387354 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.387366 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.490027 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.490085 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.490095 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.490110 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.490121 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.593004 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.593061 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.593079 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.593102 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.593119 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.695621 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.695657 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.695667 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.695683 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.695692 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.716121 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:49 crc kubenswrapper[4874]: E0122 11:41:49.716255 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.765263 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 22:21:51.232342252 +0000 UTC Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.798568 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.798607 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.798622 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.798641 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.798655 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.902051 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.902122 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.902154 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.902182 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:49 crc kubenswrapper[4874]: I0122 11:41:49.902203 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:49Z","lastTransitionTime":"2026-01-22T11:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.004261 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.004298 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.004309 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.004323 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.004332 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:50Z","lastTransitionTime":"2026-01-22T11:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.106733 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.106809 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.106830 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.106866 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.106888 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:50Z","lastTransitionTime":"2026-01-22T11:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.209988 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.210032 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.210045 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.210062 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.210073 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:50Z","lastTransitionTime":"2026-01-22T11:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.313683 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.313779 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.313795 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.313818 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.313833 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:50Z","lastTransitionTime":"2026-01-22T11:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.416881 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.416929 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.416945 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.416965 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.416979 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:50Z","lastTransitionTime":"2026-01-22T11:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.518785 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.518839 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.518855 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.518877 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.518894 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:50Z","lastTransitionTime":"2026-01-22T11:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.621386 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.621462 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.621473 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.621487 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.621496 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:50Z","lastTransitionTime":"2026-01-22T11:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.715671 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.715778 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.715853 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:50 crc kubenswrapper[4874]: E0122 11:41:50.716052 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:50 crc kubenswrapper[4874]: E0122 11:41:50.716245 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:50 crc kubenswrapper[4874]: E0122 11:41:50.716373 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.724065 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.724114 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.724128 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.724155 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.724172 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:50Z","lastTransitionTime":"2026-01-22T11:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.765930 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 18:19:26.525062654 +0000 UTC Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.827101 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.827142 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.827151 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.827169 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.827180 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:50Z","lastTransitionTime":"2026-01-22T11:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.930075 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.930143 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.930165 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.930189 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:50 crc kubenswrapper[4874]: I0122 11:41:50.930207 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:50Z","lastTransitionTime":"2026-01-22T11:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.033067 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.033126 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.033145 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.033168 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.033185 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:51Z","lastTransitionTime":"2026-01-22T11:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.136379 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.136487 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.136508 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.136531 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.136548 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:51Z","lastTransitionTime":"2026-01-22T11:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.238647 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.238696 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.238706 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.238721 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.238730 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:51Z","lastTransitionTime":"2026-01-22T11:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.341221 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.341571 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.341689 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.341782 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.341873 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:51Z","lastTransitionTime":"2026-01-22T11:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.444718 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.444783 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.444808 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.444838 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.444860 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:51Z","lastTransitionTime":"2026-01-22T11:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.547016 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.547315 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.547525 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.547669 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.547833 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:51Z","lastTransitionTime":"2026-01-22T11:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.650358 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.651172 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.651330 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.651529 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.651675 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:51Z","lastTransitionTime":"2026-01-22T11:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.715124 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:51 crc kubenswrapper[4874]: E0122 11:41:51.715302 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.754657 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.754959 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.755137 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.755300 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.755464 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:51Z","lastTransitionTime":"2026-01-22T11:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.766122 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 05:45:30.990893467 +0000 UTC Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.858045 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.858298 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.858513 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.858660 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.858790 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:51Z","lastTransitionTime":"2026-01-22T11:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.962275 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.962340 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.962362 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.962386 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:51 crc kubenswrapper[4874]: I0122 11:41:51.962437 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:51Z","lastTransitionTime":"2026-01-22T11:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.065511 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.065883 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.066087 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.066295 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.066531 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:52Z","lastTransitionTime":"2026-01-22T11:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.168965 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.169024 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.169043 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.169069 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.169088 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:52Z","lastTransitionTime":"2026-01-22T11:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.271857 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.271904 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.271915 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.271929 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.271940 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:52Z","lastTransitionTime":"2026-01-22T11:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.374022 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.374065 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.374075 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.374089 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.374099 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:52Z","lastTransitionTime":"2026-01-22T11:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.476684 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.476734 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.476752 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.476789 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.476805 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:52Z","lastTransitionTime":"2026-01-22T11:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.579112 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.579173 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.579194 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.579218 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.579235 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:52Z","lastTransitionTime":"2026-01-22T11:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.680989 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.681056 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.681079 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.681104 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.681122 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:52Z","lastTransitionTime":"2026-01-22T11:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.715999 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.716078 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.716041 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:52 crc kubenswrapper[4874]: E0122 11:41:52.716201 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:52 crc kubenswrapper[4874]: E0122 11:41:52.716480 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:52 crc kubenswrapper[4874]: E0122 11:41:52.716467 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.717624 4874 scope.go:117] "RemoveContainer" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:41:52 crc kubenswrapper[4874]: E0122 11:41:52.717872 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.766922 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 08:05:20.240315663 +0000 UTC Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.783834 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.783885 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.783898 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.783917 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.783929 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:52Z","lastTransitionTime":"2026-01-22T11:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.886766 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.886842 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.886865 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.886896 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.886916 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:52Z","lastTransitionTime":"2026-01-22T11:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.989068 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.989102 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.989112 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.989127 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:52 crc kubenswrapper[4874]: I0122 11:41:52.989135 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:52Z","lastTransitionTime":"2026-01-22T11:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.092187 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.092262 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.092285 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.092312 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.092333 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:53Z","lastTransitionTime":"2026-01-22T11:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.195123 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.195163 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.195175 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.195191 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.195204 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:53Z","lastTransitionTime":"2026-01-22T11:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.297276 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.297312 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.297326 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.297342 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.297353 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:53Z","lastTransitionTime":"2026-01-22T11:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.400117 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.400152 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.400161 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.400175 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.400184 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:53Z","lastTransitionTime":"2026-01-22T11:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.503295 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.503347 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.503365 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.503387 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.503428 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:53Z","lastTransitionTime":"2026-01-22T11:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.606997 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.607081 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.607108 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.607141 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.607174 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:53Z","lastTransitionTime":"2026-01-22T11:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.710627 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.710695 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.710718 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.710745 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.710765 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:53Z","lastTransitionTime":"2026-01-22T11:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.715179 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:53 crc kubenswrapper[4874]: E0122 11:41:53.715354 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.767660 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 02:22:40.104832518 +0000 UTC Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.813370 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.813445 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.813462 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.813484 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.813499 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:53Z","lastTransitionTime":"2026-01-22T11:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.917218 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.917288 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.917311 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.917341 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:53 crc kubenswrapper[4874]: I0122 11:41:53.917363 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:53Z","lastTransitionTime":"2026-01-22T11:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.019736 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.019785 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.019800 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.019823 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.019844 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:54Z","lastTransitionTime":"2026-01-22T11:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.123285 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.123353 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.123375 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.123440 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.123462 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:54Z","lastTransitionTime":"2026-01-22T11:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.226560 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.226936 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.227018 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.227484 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.227552 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:54Z","lastTransitionTime":"2026-01-22T11:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.330543 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.330593 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.330610 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.330632 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.330649 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:54Z","lastTransitionTime":"2026-01-22T11:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.433344 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.433376 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.433384 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.433418 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.433428 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:54Z","lastTransitionTime":"2026-01-22T11:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.536060 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.536096 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.536105 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.536120 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.536131 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:54Z","lastTransitionTime":"2026-01-22T11:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.639178 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.639217 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.639229 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.639247 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.639258 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:54Z","lastTransitionTime":"2026-01-22T11:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.716102 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.716140 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:54 crc kubenswrapper[4874]: E0122 11:41:54.716294 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.716658 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:54 crc kubenswrapper[4874]: E0122 11:41:54.716804 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:54 crc kubenswrapper[4874]: E0122 11:41:54.716833 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.741711 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.741747 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.741759 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.741776 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.741789 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:54Z","lastTransitionTime":"2026-01-22T11:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.768612 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 22:44:37.820492994 +0000 UTC Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.844816 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.844884 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.844902 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.844927 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.844945 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:54Z","lastTransitionTime":"2026-01-22T11:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.947746 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.947786 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.947797 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.947811 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:54 crc kubenswrapper[4874]: I0122 11:41:54.947824 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:54Z","lastTransitionTime":"2026-01-22T11:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.051142 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.051215 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.051240 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.051269 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.051290 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:55Z","lastTransitionTime":"2026-01-22T11:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.153752 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.153814 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.153830 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.153851 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.153866 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:55Z","lastTransitionTime":"2026-01-22T11:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.256091 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.256163 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.256189 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.256252 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.256275 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:55Z","lastTransitionTime":"2026-01-22T11:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.359572 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.359617 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.359629 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.359645 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.359656 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:55Z","lastTransitionTime":"2026-01-22T11:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.463481 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.463513 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.463523 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.463539 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.463552 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:55Z","lastTransitionTime":"2026-01-22T11:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.565781 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.565817 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.565826 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.565839 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.565848 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:55Z","lastTransitionTime":"2026-01-22T11:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.667919 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.667949 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.667958 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.667973 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.667984 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:55Z","lastTransitionTime":"2026-01-22T11:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.715963 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:55 crc kubenswrapper[4874]: E0122 11:41:55.716151 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.769073 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 06:11:27.785468592 +0000 UTC Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.770935 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.770962 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.770970 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.770983 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.770992 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:55Z","lastTransitionTime":"2026-01-22T11:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.873023 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.873111 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.873134 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.873164 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.873186 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:55Z","lastTransitionTime":"2026-01-22T11:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.975692 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.975768 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.975794 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.975823 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:55 crc kubenswrapper[4874]: I0122 11:41:55.975845 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:55Z","lastTransitionTime":"2026-01-22T11:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.078598 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.078642 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.078663 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.078682 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.078696 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:56Z","lastTransitionTime":"2026-01-22T11:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.180676 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.180757 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.180785 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.180828 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.180857 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:56Z","lastTransitionTime":"2026-01-22T11:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.283189 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.283236 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.283252 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.283272 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.283287 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:56Z","lastTransitionTime":"2026-01-22T11:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.389210 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.389840 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.389860 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.389881 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.389894 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:56Z","lastTransitionTime":"2026-01-22T11:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.493215 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.493265 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.493274 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.493292 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.493303 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:56Z","lastTransitionTime":"2026-01-22T11:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.596350 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.596451 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.596471 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.596499 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.596516 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:56Z","lastTransitionTime":"2026-01-22T11:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.699431 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.699481 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.699493 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.699509 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.699522 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:56Z","lastTransitionTime":"2026-01-22T11:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.716145 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.717259 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.717358 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:56 crc kubenswrapper[4874]: E0122 11:41:56.717481 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:56 crc kubenswrapper[4874]: E0122 11:41:56.717579 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:56 crc kubenswrapper[4874]: E0122 11:41:56.717727 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.732420 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d78bedba-59c5-4a8a-89c2-4414b15e80a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"lient-ca-file\\\\nI0122 11:40:44.444174 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0122 11:40:44.444187 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0122 11:40:44.444220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0122 11:40:44.444256 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0122 11:40:44.444339 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769082028\\\\\\\\\\\\\\\" (2026-01-22 11:40:27 +0000 UTC to 2026-02-21 11:40:28 +0000 UTC (now=2026-01-22 11:40:44.444302798 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444371 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-701345747/tls.crt::/tmp/serving-cert-701345747/tls.key\\\\\\\"\\\\nI0122 11:40:44.444515 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769082039\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769082039\\\\\\\\\\\\\\\" (2026-01-22 10:40:38 +0000 UTC to 2027-01-22 10:40:38 +0000 UTC (now=2026-01-22 11:40:44.444487504 +0000 UTC))\\\\\\\"\\\\nI0122 11:40:44.444544 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0122 11:40:44.444567 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0122 11:40:44.444571 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 11:40:44.444598 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0122 11:40:44.447070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.754576 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0047f17e-2ad9-4a43-84e3-9a40551de219\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a2f2d27202a2d969dccaf2873952a87458d6493e0349a5f9d57d34c01122496\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5677adf9234c26630e4d123a298a3b5b78023b99e2440aedb6ba3d73947258b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4dc4e9e4e97f59e7746733489f35ede2039b0a79334dae080ffce466f69b5e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd9d25eaf32a4b740ebdfa226946fff1c0e7117a837ccdf35db16226c33ef7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc3d32eccb2d3485d6d3cdda4602d0c52124a9e7358ce0f8545c0df8d3bc572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3a57545c435ec01d831ca3e781602465e8888f68010adb602176ee78bcb78f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9991227fae7d5e5b98375d16ef781c350d7c884df8875c48a3d3811031241b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa8313d3c4ced0c9a00c8c60b930f8fd257455bbc79e228fff3c0735d60b6a90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.769750 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 14:04:40.474966684 +0000 UTC Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.769894 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f82d39b1-4b1a-416e-af98-12acfb504203\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb247eedc7bd6d1590960a8f923c7a967873820b48176afaa52cfa2a5d8f863a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1def1c12d8ce72f4f05db7ba4fb7322bf2df6de33fbec65dda064988256461a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12b7c1415511ab55581943924cdd0e2fcd0dd85e6c2e820110095a843a7ff00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.781917 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.791754 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5451fbab-ebad-42e7-bb80-f94bad10d571\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmzdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T11:40:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lr2vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.802785 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.802832 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.802842 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.802856 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.802865 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:56Z","lastTransitionTime":"2026-01-22T11:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.804077 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f3995504d35f199ad3862c187f002d8c17f8700bead094aa251247f29b87ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.817791 4874 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T11:40:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea295b271c741c5d2ad965686077935d0b12cd4108544715b905fed81d65c453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddc55a83ed3e6e3cb1c06641bd5523431e37d70f2fdb9c60f854d4b9d7e38b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T11:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T11:41:56Z is after 2025-08-24T17:21:41Z" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.871915 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=17.871890507 podStartE2EDuration="17.871890507s" podCreationTimestamp="2026-01-22 11:41:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:41:56.85749038 +0000 UTC m=+90.702561480" watchObservedRunningTime="2026-01-22 11:41:56.871890507 +0000 UTC m=+90.716961577" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.886857 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=41.886831581 podStartE2EDuration="41.886831581s" podCreationTimestamp="2026-01-22 11:41:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:41:56.872472816 +0000 UTC m=+90.717543906" watchObservedRunningTime="2026-01-22 11:41:56.886831581 +0000 UTC m=+90.731902661" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.899532 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-krrtc" podStartSLOduration=72.899513814 podStartE2EDuration="1m12.899513814s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:41:56.898965917 +0000 UTC m=+90.744036987" watchObservedRunningTime="2026-01-22 11:41:56.899513814 +0000 UTC m=+90.744584884" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.906458 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.906497 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.906508 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.906526 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.906536 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:56Z","lastTransitionTime":"2026-01-22T11:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.917263 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-pdb2m" podStartSLOduration=72.917243008 podStartE2EDuration="1m12.917243008s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:41:56.917170105 +0000 UTC m=+90.762241185" watchObservedRunningTime="2026-01-22 11:41:56.917243008 +0000 UTC m=+90.762314078" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.942761 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-q2rnk" podStartSLOduration=72.942737526 podStartE2EDuration="1m12.942737526s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:41:56.942603653 +0000 UTC m=+90.787674743" watchObservedRunningTime="2026-01-22 11:41:56.942737526 +0000 UTC m=+90.787808596" Jan 22 11:41:56 crc kubenswrapper[4874]: I0122 11:41:56.964498 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5vd5" podStartSLOduration=71.964476987 podStartE2EDuration="1m11.964476987s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:41:56.964133646 +0000 UTC m=+90.809204726" watchObservedRunningTime="2026-01-22 11:41:56.964476987 +0000 UTC m=+90.809548067" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.008713 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-prbck" podStartSLOduration=73.008652 podStartE2EDuration="1m13.008652s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:41:57.008449784 +0000 UTC m=+90.853520854" watchObservedRunningTime="2026-01-22 11:41:57.008652 +0000 UTC m=+90.853723070" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.009096 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.009129 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.009143 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.009166 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.009179 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:57Z","lastTransitionTime":"2026-01-22T11:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.023602 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podStartSLOduration=73.023576084 podStartE2EDuration="1m13.023576084s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:41:57.022874972 +0000 UTC m=+90.867946042" watchObservedRunningTime="2026-01-22 11:41:57.023576084 +0000 UTC m=+90.868647154" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.111424 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.111474 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.111483 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.111499 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.111510 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:57Z","lastTransitionTime":"2026-01-22T11:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.213635 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.213690 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.213705 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.213727 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.213745 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:57Z","lastTransitionTime":"2026-01-22T11:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.316305 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.316356 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.316370 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.316391 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.316429 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:57Z","lastTransitionTime":"2026-01-22T11:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.419003 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.419040 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.419051 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.419068 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.419078 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:57Z","lastTransitionTime":"2026-01-22T11:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.521802 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.522094 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.522165 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.522242 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.522314 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:57Z","lastTransitionTime":"2026-01-22T11:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.625562 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.625626 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.625649 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.625674 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.625692 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:57Z","lastTransitionTime":"2026-01-22T11:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.715967 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:57 crc kubenswrapper[4874]: E0122 11:41:57.716170 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.727954 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.728003 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.728011 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.728026 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.728036 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:57Z","lastTransitionTime":"2026-01-22T11:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.770148 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 04:11:22.252468638 +0000 UTC Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.830501 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.830563 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.830580 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.830607 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.830624 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:57Z","lastTransitionTime":"2026-01-22T11:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.933823 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.933873 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.933889 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.933914 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:57 crc kubenswrapper[4874]: I0122 11:41:57.933933 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:57Z","lastTransitionTime":"2026-01-22T11:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.036898 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.037010 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.037036 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.037070 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.037097 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:58Z","lastTransitionTime":"2026-01-22T11:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.139124 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.139170 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.139181 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.139197 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.139208 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:58Z","lastTransitionTime":"2026-01-22T11:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.242261 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.242295 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.242307 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.242329 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.242340 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:58Z","lastTransitionTime":"2026-01-22T11:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.345313 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.345379 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.345425 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.345451 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.345468 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:58Z","lastTransitionTime":"2026-01-22T11:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.448294 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.448343 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.448352 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.448369 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.448380 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:58Z","lastTransitionTime":"2026-01-22T11:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.550629 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.550685 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.550701 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.550720 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.550736 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:58Z","lastTransitionTime":"2026-01-22T11:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.653166 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.653207 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.653221 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.653236 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.653247 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:58Z","lastTransitionTime":"2026-01-22T11:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.716208 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.716270 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:41:58 crc kubenswrapper[4874]: E0122 11:41:58.716329 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.716208 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:41:58 crc kubenswrapper[4874]: E0122 11:41:58.716389 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:41:58 crc kubenswrapper[4874]: E0122 11:41:58.716689 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.755960 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.756019 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.756041 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.756065 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.756082 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:58Z","lastTransitionTime":"2026-01-22T11:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.770380 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 18:01:29.590806248 +0000 UTC Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.859151 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.859202 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.859218 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.859239 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.859256 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:58Z","lastTransitionTime":"2026-01-22T11:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.962125 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.962196 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.962210 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.962238 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:58 crc kubenswrapper[4874]: I0122 11:41:58.962261 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:58Z","lastTransitionTime":"2026-01-22T11:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.065661 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.065734 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.065755 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.065782 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.065799 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:59Z","lastTransitionTime":"2026-01-22T11:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.169439 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.169525 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.169546 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.169568 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.169588 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:59Z","lastTransitionTime":"2026-01-22T11:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.272599 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.273016 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.273221 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.273490 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.273695 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:59Z","lastTransitionTime":"2026-01-22T11:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.286317 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.286360 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.286377 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.286449 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.286476 4874 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T11:41:59Z","lastTransitionTime":"2026-01-22T11:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.345258 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw"] Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.345996 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.348078 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.348547 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.348922 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.349622 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.396306 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=73.39628937 podStartE2EDuration="1m13.39628937s" podCreationTimestamp="2026-01-22 11:40:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:41:59.39534681 +0000 UTC m=+93.240417870" watchObservedRunningTime="2026-01-22 11:41:59.39628937 +0000 UTC m=+93.241360440" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.396523 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=75.396518157 podStartE2EDuration="1m15.396518157s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:41:59.36698123 +0000 UTC m=+93.212052340" watchObservedRunningTime="2026-01-22 11:41:59.396518157 +0000 UTC m=+93.241589227" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.409464 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=67.409447228 podStartE2EDuration="1m7.409447228s" podCreationTimestamp="2026-01-22 11:40:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:41:59.409295093 +0000 UTC m=+93.254366173" watchObservedRunningTime="2026-01-22 11:41:59.409447228 +0000 UTC m=+93.254518298" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.445794 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/85f182db-0073-40ce-ae9b-94bbbef42279-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.445857 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/85f182db-0073-40ce-ae9b-94bbbef42279-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.445875 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/85f182db-0073-40ce-ae9b-94bbbef42279-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.445907 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85f182db-0073-40ce-ae9b-94bbbef42279-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.445921 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/85f182db-0073-40ce-ae9b-94bbbef42279-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.546473 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85f182db-0073-40ce-ae9b-94bbbef42279-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.546529 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/85f182db-0073-40ce-ae9b-94bbbef42279-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.546576 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/85f182db-0073-40ce-ae9b-94bbbef42279-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.546656 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/85f182db-0073-40ce-ae9b-94bbbef42279-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.546684 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/85f182db-0073-40ce-ae9b-94bbbef42279-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.546748 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/85f182db-0073-40ce-ae9b-94bbbef42279-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.546767 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/85f182db-0073-40ce-ae9b-94bbbef42279-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.547468 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/85f182db-0073-40ce-ae9b-94bbbef42279-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.561189 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85f182db-0073-40ce-ae9b-94bbbef42279-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.565429 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/85f182db-0073-40ce-ae9b-94bbbef42279-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qfxsw\" (UID: \"85f182db-0073-40ce-ae9b-94bbbef42279\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.670674 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" Jan 22 11:41:59 crc kubenswrapper[4874]: W0122 11:41:59.687472 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85f182db_0073_40ce_ae9b_94bbbef42279.slice/crio-6784cd319e5c3ab88125d0a600b509ee3650d8d47aea2791f95196f3f207a8ec WatchSource:0}: Error finding container 6784cd319e5c3ab88125d0a600b509ee3650d8d47aea2791f95196f3f207a8ec: Status 404 returned error can't find the container with id 6784cd319e5c3ab88125d0a600b509ee3650d8d47aea2791f95196f3f207a8ec Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.715511 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:41:59 crc kubenswrapper[4874]: E0122 11:41:59.715687 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.771045 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 19:00:35.117662431 +0000 UTC Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.771130 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 22 11:41:59 crc kubenswrapper[4874]: I0122 11:41:59.779139 4874 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 22 11:42:00 crc kubenswrapper[4874]: I0122 11:42:00.555161 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" event={"ID":"85f182db-0073-40ce-ae9b-94bbbef42279","Type":"ContainerStarted","Data":"e6966912417cfcda36c2d50d4aad5c911d1c1773b45ce73ea88d890e1bc29126"} Jan 22 11:42:00 crc kubenswrapper[4874]: I0122 11:42:00.555232 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" event={"ID":"85f182db-0073-40ce-ae9b-94bbbef42279","Type":"ContainerStarted","Data":"6784cd319e5c3ab88125d0a600b509ee3650d8d47aea2791f95196f3f207a8ec"} Jan 22 11:42:00 crc kubenswrapper[4874]: I0122 11:42:00.577188 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qfxsw" podStartSLOduration=76.577161009 podStartE2EDuration="1m16.577161009s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:00.57623817 +0000 UTC m=+94.421309270" watchObservedRunningTime="2026-01-22 11:42:00.577161009 +0000 UTC m=+94.422232109" Jan 22 11:42:00 crc kubenswrapper[4874]: I0122 11:42:00.715862 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:00 crc kubenswrapper[4874]: I0122 11:42:00.715968 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:00 crc kubenswrapper[4874]: E0122 11:42:00.716129 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:00 crc kubenswrapper[4874]: E0122 11:42:00.716361 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:00 crc kubenswrapper[4874]: I0122 11:42:00.717275 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:00 crc kubenswrapper[4874]: E0122 11:42:00.717514 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:01 crc kubenswrapper[4874]: I0122 11:42:01.715803 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:01 crc kubenswrapper[4874]: E0122 11:42:01.715959 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:02 crc kubenswrapper[4874]: I0122 11:42:02.480439 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:02 crc kubenswrapper[4874]: E0122 11:42:02.480716 4874 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:42:02 crc kubenswrapper[4874]: E0122 11:42:02.480972 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs podName:5451fbab-ebad-42e7-bb80-f94bad10d571 nodeName:}" failed. No retries permitted until 2026-01-22 11:43:06.480952295 +0000 UTC m=+160.326023365 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs") pod "network-metrics-daemon-lr2vd" (UID: "5451fbab-ebad-42e7-bb80-f94bad10d571") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 11:42:02 crc kubenswrapper[4874]: I0122 11:42:02.715799 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:02 crc kubenswrapper[4874]: E0122 11:42:02.716287 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:02 crc kubenswrapper[4874]: I0122 11:42:02.715894 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:02 crc kubenswrapper[4874]: E0122 11:42:02.717030 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:02 crc kubenswrapper[4874]: I0122 11:42:02.715885 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:02 crc kubenswrapper[4874]: E0122 11:42:02.717552 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:03 crc kubenswrapper[4874]: I0122 11:42:03.715316 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:03 crc kubenswrapper[4874]: E0122 11:42:03.715623 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:04 crc kubenswrapper[4874]: I0122 11:42:04.716151 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:04 crc kubenswrapper[4874]: I0122 11:42:04.716207 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:04 crc kubenswrapper[4874]: E0122 11:42:04.716272 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:04 crc kubenswrapper[4874]: E0122 11:42:04.716373 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:04 crc kubenswrapper[4874]: I0122 11:42:04.716559 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:04 crc kubenswrapper[4874]: E0122 11:42:04.716691 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:04 crc kubenswrapper[4874]: I0122 11:42:04.717695 4874 scope.go:117] "RemoveContainer" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:42:04 crc kubenswrapper[4874]: E0122 11:42:04.717987 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" Jan 22 11:42:05 crc kubenswrapper[4874]: I0122 11:42:05.714980 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:05 crc kubenswrapper[4874]: E0122 11:42:05.715094 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:06 crc kubenswrapper[4874]: I0122 11:42:06.716213 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:06 crc kubenswrapper[4874]: I0122 11:42:06.716224 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:06 crc kubenswrapper[4874]: I0122 11:42:06.716294 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:06 crc kubenswrapper[4874]: E0122 11:42:06.717512 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:06 crc kubenswrapper[4874]: E0122 11:42:06.717735 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:06 crc kubenswrapper[4874]: E0122 11:42:06.717931 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:07 crc kubenswrapper[4874]: I0122 11:42:07.716042 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:07 crc kubenswrapper[4874]: E0122 11:42:07.716280 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:08 crc kubenswrapper[4874]: I0122 11:42:08.716047 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:08 crc kubenswrapper[4874]: I0122 11:42:08.716117 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:08 crc kubenswrapper[4874]: I0122 11:42:08.716055 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:08 crc kubenswrapper[4874]: E0122 11:42:08.716191 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:08 crc kubenswrapper[4874]: E0122 11:42:08.716303 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:08 crc kubenswrapper[4874]: E0122 11:42:08.716467 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:09 crc kubenswrapper[4874]: I0122 11:42:09.715608 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:09 crc kubenswrapper[4874]: E0122 11:42:09.715932 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:10 crc kubenswrapper[4874]: I0122 11:42:10.715856 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:10 crc kubenswrapper[4874]: I0122 11:42:10.715896 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:10 crc kubenswrapper[4874]: E0122 11:42:10.716017 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:10 crc kubenswrapper[4874]: I0122 11:42:10.715856 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:10 crc kubenswrapper[4874]: E0122 11:42:10.716114 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:10 crc kubenswrapper[4874]: E0122 11:42:10.716174 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:11 crc kubenswrapper[4874]: I0122 11:42:11.715175 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:11 crc kubenswrapper[4874]: E0122 11:42:11.715349 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:12 crc kubenswrapper[4874]: I0122 11:42:12.715498 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:12 crc kubenswrapper[4874]: E0122 11:42:12.715661 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:12 crc kubenswrapper[4874]: I0122 11:42:12.715857 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:12 crc kubenswrapper[4874]: I0122 11:42:12.715531 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:12 crc kubenswrapper[4874]: E0122 11:42:12.716020 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:12 crc kubenswrapper[4874]: E0122 11:42:12.716133 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:13 crc kubenswrapper[4874]: I0122 11:42:13.715725 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:13 crc kubenswrapper[4874]: E0122 11:42:13.715912 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:14 crc kubenswrapper[4874]: I0122 11:42:14.715314 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:14 crc kubenswrapper[4874]: I0122 11:42:14.715342 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:14 crc kubenswrapper[4874]: E0122 11:42:14.715515 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:14 crc kubenswrapper[4874]: I0122 11:42:14.715570 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:14 crc kubenswrapper[4874]: E0122 11:42:14.715754 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:14 crc kubenswrapper[4874]: E0122 11:42:14.715819 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:15 crc kubenswrapper[4874]: I0122 11:42:15.715877 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:15 crc kubenswrapper[4874]: E0122 11:42:15.716079 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:16 crc kubenswrapper[4874]: I0122 11:42:16.715491 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:16 crc kubenswrapper[4874]: I0122 11:42:16.715594 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:16 crc kubenswrapper[4874]: E0122 11:42:16.718381 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:16 crc kubenswrapper[4874]: I0122 11:42:16.718477 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:16 crc kubenswrapper[4874]: E0122 11:42:16.718743 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:16 crc kubenswrapper[4874]: E0122 11:42:16.718621 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:17 crc kubenswrapper[4874]: I0122 11:42:17.619666 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/1.log" Jan 22 11:42:17 crc kubenswrapper[4874]: I0122 11:42:17.620662 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/0.log" Jan 22 11:42:17 crc kubenswrapper[4874]: I0122 11:42:17.620708 4874 generic.go:334] "Generic (PLEG): container finished" podID="977746b5-ac1b-4b6e-bdbc-ddd90225e68c" containerID="cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8" exitCode=1 Jan 22 11:42:17 crc kubenswrapper[4874]: I0122 11:42:17.620776 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-krrtc" event={"ID":"977746b5-ac1b-4b6e-bdbc-ddd90225e68c","Type":"ContainerDied","Data":"cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8"} Jan 22 11:42:17 crc kubenswrapper[4874]: I0122 11:42:17.620999 4874 scope.go:117] "RemoveContainer" containerID="600d18ed6bc497772f5a8a2f70412d8e845c43e6f54c9f30666f362ab94aacbf" Jan 22 11:42:17 crc kubenswrapper[4874]: I0122 11:42:17.622180 4874 scope.go:117] "RemoveContainer" containerID="cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8" Jan 22 11:42:17 crc kubenswrapper[4874]: E0122 11:42:17.622481 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-krrtc_openshift-multus(977746b5-ac1b-4b6e-bdbc-ddd90225e68c)\"" pod="openshift-multus/multus-krrtc" podUID="977746b5-ac1b-4b6e-bdbc-ddd90225e68c" Jan 22 11:42:17 crc kubenswrapper[4874]: I0122 11:42:17.716191 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:17 crc kubenswrapper[4874]: E0122 11:42:17.716483 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:18 crc kubenswrapper[4874]: I0122 11:42:18.624769 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/1.log" Jan 22 11:42:18 crc kubenswrapper[4874]: I0122 11:42:18.715999 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:18 crc kubenswrapper[4874]: I0122 11:42:18.716020 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:18 crc kubenswrapper[4874]: I0122 11:42:18.716136 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:18 crc kubenswrapper[4874]: E0122 11:42:18.716252 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:18 crc kubenswrapper[4874]: E0122 11:42:18.716442 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:18 crc kubenswrapper[4874]: E0122 11:42:18.716580 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:18 crc kubenswrapper[4874]: I0122 11:42:18.717307 4874 scope.go:117] "RemoveContainer" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:42:18 crc kubenswrapper[4874]: E0122 11:42:18.717512 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6tmll_openshift-ovn-kubernetes(642d0ca0-2e0f-4b69-9484-a63d0a01f8a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" Jan 22 11:42:19 crc kubenswrapper[4874]: I0122 11:42:19.715536 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:19 crc kubenswrapper[4874]: E0122 11:42:19.716205 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:20 crc kubenswrapper[4874]: I0122 11:42:20.717473 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:20 crc kubenswrapper[4874]: I0122 11:42:20.717631 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:20 crc kubenswrapper[4874]: E0122 11:42:20.717652 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:20 crc kubenswrapper[4874]: I0122 11:42:20.717937 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:20 crc kubenswrapper[4874]: E0122 11:42:20.718111 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:20 crc kubenswrapper[4874]: E0122 11:42:20.717868 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:21 crc kubenswrapper[4874]: I0122 11:42:21.715524 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:21 crc kubenswrapper[4874]: E0122 11:42:21.715694 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:22 crc kubenswrapper[4874]: I0122 11:42:22.715665 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:22 crc kubenswrapper[4874]: E0122 11:42:22.715783 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:22 crc kubenswrapper[4874]: I0122 11:42:22.715766 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:22 crc kubenswrapper[4874]: I0122 11:42:22.715936 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:22 crc kubenswrapper[4874]: E0122 11:42:22.716136 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:22 crc kubenswrapper[4874]: E0122 11:42:22.716490 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:23 crc kubenswrapper[4874]: I0122 11:42:23.715841 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:23 crc kubenswrapper[4874]: E0122 11:42:23.716139 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:24 crc kubenswrapper[4874]: I0122 11:42:24.716091 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:24 crc kubenswrapper[4874]: I0122 11:42:24.716207 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:24 crc kubenswrapper[4874]: I0122 11:42:24.716091 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:24 crc kubenswrapper[4874]: E0122 11:42:24.716325 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:24 crc kubenswrapper[4874]: E0122 11:42:24.716451 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:24 crc kubenswrapper[4874]: E0122 11:42:24.716531 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:25 crc kubenswrapper[4874]: I0122 11:42:25.716106 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:25 crc kubenswrapper[4874]: E0122 11:42:25.716752 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:26 crc kubenswrapper[4874]: I0122 11:42:26.715871 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:26 crc kubenswrapper[4874]: I0122 11:42:26.715998 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:26 crc kubenswrapper[4874]: E0122 11:42:26.716102 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:26 crc kubenswrapper[4874]: I0122 11:42:26.716175 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:26 crc kubenswrapper[4874]: E0122 11:42:26.718434 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:26 crc kubenswrapper[4874]: E0122 11:42:26.718623 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:26 crc kubenswrapper[4874]: E0122 11:42:26.732648 4874 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 22 11:42:26 crc kubenswrapper[4874]: E0122 11:42:26.818091 4874 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 22 11:42:27 crc kubenswrapper[4874]: I0122 11:42:27.715978 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:27 crc kubenswrapper[4874]: E0122 11:42:27.716177 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:28 crc kubenswrapper[4874]: I0122 11:42:28.716082 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:28 crc kubenswrapper[4874]: E0122 11:42:28.716279 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:28 crc kubenswrapper[4874]: I0122 11:42:28.717652 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:28 crc kubenswrapper[4874]: E0122 11:42:28.717804 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:28 crc kubenswrapper[4874]: I0122 11:42:28.716042 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:28 crc kubenswrapper[4874]: E0122 11:42:28.719728 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:29 crc kubenswrapper[4874]: I0122 11:42:29.716016 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:29 crc kubenswrapper[4874]: E0122 11:42:29.716652 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:29 crc kubenswrapper[4874]: I0122 11:42:29.718108 4874 scope.go:117] "RemoveContainer" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:42:30 crc kubenswrapper[4874]: I0122 11:42:30.624710 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lr2vd"] Jan 22 11:42:30 crc kubenswrapper[4874]: I0122 11:42:30.671315 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/3.log" Jan 22 11:42:30 crc kubenswrapper[4874]: I0122 11:42:30.673871 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerStarted","Data":"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161"} Jan 22 11:42:30 crc kubenswrapper[4874]: I0122 11:42:30.673911 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:30 crc kubenswrapper[4874]: E0122 11:42:30.674028 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:30 crc kubenswrapper[4874]: I0122 11:42:30.674256 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:42:30 crc kubenswrapper[4874]: I0122 11:42:30.704874 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podStartSLOduration=106.704852279 podStartE2EDuration="1m46.704852279s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:30.704609281 +0000 UTC m=+124.549680371" watchObservedRunningTime="2026-01-22 11:42:30.704852279 +0000 UTC m=+124.549923349" Jan 22 11:42:30 crc kubenswrapper[4874]: I0122 11:42:30.715211 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:30 crc kubenswrapper[4874]: I0122 11:42:30.715257 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:30 crc kubenswrapper[4874]: I0122 11:42:30.715292 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:30 crc kubenswrapper[4874]: E0122 11:42:30.715343 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:30 crc kubenswrapper[4874]: E0122 11:42:30.715545 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:30 crc kubenswrapper[4874]: E0122 11:42:30.715750 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:31 crc kubenswrapper[4874]: E0122 11:42:31.820112 4874 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 22 11:42:32 crc kubenswrapper[4874]: I0122 11:42:32.715275 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:32 crc kubenswrapper[4874]: I0122 11:42:32.715343 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:32 crc kubenswrapper[4874]: E0122 11:42:32.715462 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:32 crc kubenswrapper[4874]: I0122 11:42:32.715275 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:32 crc kubenswrapper[4874]: I0122 11:42:32.715539 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:32 crc kubenswrapper[4874]: E0122 11:42:32.715774 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:32 crc kubenswrapper[4874]: E0122 11:42:32.716185 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:32 crc kubenswrapper[4874]: I0122 11:42:32.716323 4874 scope.go:117] "RemoveContainer" containerID="cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8" Jan 22 11:42:32 crc kubenswrapper[4874]: E0122 11:42:32.716363 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:33 crc kubenswrapper[4874]: I0122 11:42:33.687249 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/1.log" Jan 22 11:42:33 crc kubenswrapper[4874]: I0122 11:42:33.687337 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-krrtc" event={"ID":"977746b5-ac1b-4b6e-bdbc-ddd90225e68c","Type":"ContainerStarted","Data":"55eeb9abd8c425711e374c107c22ec24d1741880327f226d7db5e06d67925630"} Jan 22 11:42:34 crc kubenswrapper[4874]: I0122 11:42:34.716239 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:34 crc kubenswrapper[4874]: I0122 11:42:34.716278 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:34 crc kubenswrapper[4874]: I0122 11:42:34.716369 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:34 crc kubenswrapper[4874]: E0122 11:42:34.716451 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:34 crc kubenswrapper[4874]: I0122 11:42:34.716547 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:34 crc kubenswrapper[4874]: E0122 11:42:34.717001 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:34 crc kubenswrapper[4874]: E0122 11:42:34.717274 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:34 crc kubenswrapper[4874]: E0122 11:42:34.717538 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:36 crc kubenswrapper[4874]: I0122 11:42:36.716029 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:36 crc kubenswrapper[4874]: I0122 11:42:36.716087 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:36 crc kubenswrapper[4874]: I0122 11:42:36.716177 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:36 crc kubenswrapper[4874]: I0122 11:42:36.716252 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:36 crc kubenswrapper[4874]: E0122 11:42:36.718075 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 11:42:36 crc kubenswrapper[4874]: E0122 11:42:36.718241 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 11:42:36 crc kubenswrapper[4874]: E0122 11:42:36.718587 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 11:42:36 crc kubenswrapper[4874]: E0122 11:42:36.718632 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lr2vd" podUID="5451fbab-ebad-42e7-bb80-f94bad10d571" Jan 22 11:42:38 crc kubenswrapper[4874]: I0122 11:42:38.716149 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:38 crc kubenswrapper[4874]: I0122 11:42:38.716215 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:38 crc kubenswrapper[4874]: I0122 11:42:38.716959 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:42:38 crc kubenswrapper[4874]: I0122 11:42:38.717307 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:38 crc kubenswrapper[4874]: I0122 11:42:38.718746 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 22 11:42:38 crc kubenswrapper[4874]: I0122 11:42:38.719530 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 22 11:42:38 crc kubenswrapper[4874]: I0122 11:42:38.719716 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 22 11:42:38 crc kubenswrapper[4874]: I0122 11:42:38.719893 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 22 11:42:38 crc kubenswrapper[4874]: I0122 11:42:38.720259 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 22 11:42:38 crc kubenswrapper[4874]: I0122 11:42:38.722111 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.498520 4874 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.544806 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-swcnq"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.545216 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.547375 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.547804 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.548766 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.549823 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wmwcz"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.549883 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.550136 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.550897 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.551091 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-869fn"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.551435 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.552375 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.552609 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.552969 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.559126 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.566809 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.568257 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.570126 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.571207 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.597280 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.598818 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.599307 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.599515 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.599637 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.600502 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.600517 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.600832 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.600948 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.600966 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.601116 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.601247 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.601520 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.601635 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.602123 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.602203 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.603914 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.606620 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.607526 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mftbv"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.607875 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.608111 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.608197 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-jgw4c"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.608248 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.608382 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.608554 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-jgw4c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.609027 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.609538 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.609674 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.609889 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610355 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5fb7b4d4-0441-43c4-9596-4d38b369d661-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610390 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610432 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5fb7b4d4-0441-43c4-9596-4d38b369d661-service-ca-bundle\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610456 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-config\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610478 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610499 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9bec261f-fd1e-44f7-a402-cae34f722b6c-images\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610519 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca44490a-7fc8-478d-a6e2-670f49816b81-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610552 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca44490a-7fc8-478d-a6e2-670f49816b81-encryption-config\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610583 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610603 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5464a23-ec80-4717-bfe0-6efeab811853-serving-cert\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610624 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca44490a-7fc8-478d-a6e2-670f49816b81-etcd-client\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610645 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca44490a-7fc8-478d-a6e2-670f49816b81-audit-policies\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610672 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca44490a-7fc8-478d-a6e2-670f49816b81-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610694 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkqws\" (UniqueName: \"kubernetes.io/projected/ca44490a-7fc8-478d-a6e2-670f49816b81-kube-api-access-vkqws\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610719 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9bec261f-fd1e-44f7-a402-cae34f722b6c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610740 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-machine-approver-tls\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610766 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb7b4d4-0441-43c4-9596-4d38b369d661-config\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610786 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nm6v\" (UniqueName: \"kubernetes.io/projected/f5464a23-ec80-4717-bfe0-6efeab811853-kube-api-access-2nm6v\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610806 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bec261f-fd1e-44f7-a402-cae34f722b6c-config\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610828 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-auth-proxy-config\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610847 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-client-ca\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610868 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb7b4d4-0441-43c4-9596-4d38b369d661-serving-cert\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610889 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-config\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610909 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b97ls\" (UniqueName: \"kubernetes.io/projected/9bec261f-fd1e-44f7-a402-cae34f722b6c-kube-api-access-b97ls\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610934 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca44490a-7fc8-478d-a6e2-670f49816b81-audit-dir\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610965 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca44490a-7fc8-478d-a6e2-670f49816b81-serving-cert\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.610995 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsxpl\" (UniqueName: \"kubernetes.io/projected/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-kube-api-access-tsxpl\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.611018 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nfcq\" (UniqueName: \"kubernetes.io/projected/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-kube-api-access-9nfcq\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.611040 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.611063 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7qdr\" (UniqueName: \"kubernetes.io/projected/5fb7b4d4-0441-43c4-9596-4d38b369d661-kube-api-access-x7qdr\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.611600 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.611874 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.612083 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.612095 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.613196 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.613414 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.613674 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.621843 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.621889 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.621897 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.621933 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.622199 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.624239 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.625019 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.625262 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.625439 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.625529 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.625565 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.627630 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.630485 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8j9tz"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.630882 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-cqkbr"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.631293 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.631441 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.631639 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.632705 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.637609 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.637726 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.637847 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.638061 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.641175 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.642870 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.644009 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.644297 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.644537 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.644659 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.644779 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.644971 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.655922 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.656234 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.657360 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-wws2s"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.659187 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.659891 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.660061 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.660193 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.660423 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.661292 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.661567 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.661851 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.678723 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.678886 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.679199 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.683077 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.683804 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.684144 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.684334 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.684597 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.685012 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.685205 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.685606 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.685639 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.685654 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.685735 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.686118 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.686287 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.686522 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-5cczs"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.686924 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.690214 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.690656 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.690805 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.690899 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.691160 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.690904 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.692112 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.692316 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.692540 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.692647 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.693736 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.693867 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.694215 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.694310 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.694632 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-swcnq"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.694671 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.695455 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.695842 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.696458 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.697252 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.697989 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.698170 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.699524 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.699667 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.699806 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.699920 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.701706 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.701996 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.702320 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.702355 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.702517 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.702572 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.703538 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.704428 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rx6nv"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.705173 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.705323 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jg4wj"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.705848 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.707378 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.707964 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.708493 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.708779 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.712019 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.712638 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.712905 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.713282 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.714253 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-sjhjk"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.716385 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.717330 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca44490a-7fc8-478d-a6e2-670f49816b81-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.717386 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-dir\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.717450 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2xwq\" (UniqueName: \"kubernetes.io/projected/308c337e-1e28-4e34-9ccb-8ae546eee089-kube-api-access-j2xwq\") pod \"cluster-samples-operator-665b6dd947-qfcsk\" (UID: \"308c337e-1e28-4e34-9ccb-8ae546eee089\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.717532 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kg7v\" (UniqueName: \"kubernetes.io/projected/16747491-2b7d-4cb1-841d-61c6f366cf8a-kube-api-access-2kg7v\") pod \"openshift-config-operator-7777fb866f-gv9x5\" (UID: \"16747491-2b7d-4cb1-841d-61c6f366cf8a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.717574 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/308c337e-1e28-4e34-9ccb-8ae546eee089-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qfcsk\" (UID: \"308c337e-1e28-4e34-9ccb-8ae546eee089\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.717825 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca44490a-7fc8-478d-a6e2-670f49816b81-encryption-config\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.717880 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/288e6358-c74b-4597-8968-726a31365f82-audit-dir\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.717927 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9hxd\" (UniqueName: \"kubernetes.io/projected/68dcdc32-485b-435c-81cc-43be463998bb-kube-api-access-c9hxd\") pod \"openshift-apiserver-operator-796bbdcf4f-dnrhx\" (UID: \"68dcdc32-485b-435c-81cc-43be463998bb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718102 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-policies\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718156 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718193 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5464a23-ec80-4717-bfe0-6efeab811853-serving-cert\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718235 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-service-ca\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718288 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca44490a-7fc8-478d-a6e2-670f49816b81-etcd-client\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718321 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718350 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca44490a-7fc8-478d-a6e2-670f49816b81-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718657 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca44490a-7fc8-478d-a6e2-670f49816b81-audit-policies\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718699 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-console-config\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718740 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/16747491-2b7d-4cb1-841d-61c6f366cf8a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gv9x5\" (UID: \"16747491-2b7d-4cb1-841d-61c6f366cf8a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718778 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-client-ca\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.718810 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9446a39-9776-4de8-9137-b8952d336419-service-ca-bundle\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.719135 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca44490a-7fc8-478d-a6e2-670f49816b81-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.719184 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.719228 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.719375 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh722\" (UniqueName: \"kubernetes.io/projected/304c66b8-6187-47ac-9c57-235c634eaae4-kube-api-access-sh722\") pod \"migrator-59844c95c7-rwcrv\" (UID: \"304c66b8-6187-47ac-9c57-235c634eaae4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.719431 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/288e6358-c74b-4597-8968-726a31365f82-node-pullsecrets\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.719471 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkqws\" (UniqueName: \"kubernetes.io/projected/ca44490a-7fc8-478d-a6e2-670f49816b81-kube-api-access-vkqws\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.719662 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2dad3db6-cddd-457d-8efa-908257ef7cc5-console-serving-cert\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.719691 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9jtx\" (UniqueName: \"kubernetes.io/projected/c045a558-1990-432e-9965-e918f60aba14-kube-api-access-b9jtx\") pod \"machine-config-controller-84d6567774-dprzc\" (UID: \"c045a558-1990-432e-9965-e918f60aba14\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.719722 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m29d2\" (UniqueName: \"kubernetes.io/projected/65f71c2e-ab34-4d33-905f-609555dab78c-kube-api-access-m29d2\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.719932 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj795\" (UniqueName: \"kubernetes.io/projected/3b9983ba-ed8d-4654-ba74-f25433aa7ee7-kube-api-access-gj795\") pod \"downloads-7954f5f757-jgw4c\" (UID: \"3b9983ba-ed8d-4654-ba74-f25433aa7ee7\") " pod="openshift-console/downloads-7954f5f757-jgw4c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.719984 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9bec261f-fd1e-44f7-a402-cae34f722b6c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.720026 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-machine-approver-tls\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.720071 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2dad3db6-cddd-457d-8efa-908257ef7cc5-console-oauth-config\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.720116 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rprpc\" (UniqueName: \"kubernetes.io/projected/2dad3db6-cddd-457d-8efa-908257ef7cc5-kube-api-access-rprpc\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.720315 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.720350 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-image-import-ca\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.720383 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/288e6358-c74b-4597-8968-726a31365f82-serving-cert\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.720450 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16747491-2b7d-4cb1-841d-61c6f366cf8a-serving-cert\") pod \"openshift-config-operator-7777fb866f-gv9x5\" (UID: \"16747491-2b7d-4cb1-841d-61c6f366cf8a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.720497 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.722597 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c045a558-1990-432e-9965-e918f60aba14-proxy-tls\") pod \"machine-config-controller-84d6567774-dprzc\" (UID: \"c045a558-1990-432e-9965-e918f60aba14\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.722650 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65f71c2e-ab34-4d33-905f-609555dab78c-serving-cert\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.722670 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dm8p\" (UniqueName: \"kubernetes.io/projected/88c9b59f-1809-4252-a8fc-cad965848dc0-kube-api-access-8dm8p\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.723346 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca44490a-7fc8-478d-a6e2-670f49816b81-audit-policies\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.723383 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e9446a39-9776-4de8-9137-b8952d336419-stats-auth\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.723444 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.723469 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.723490 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmkdz\" (UniqueName: \"kubernetes.io/projected/288e6358-c74b-4597-8968-726a31365f82-kube-api-access-qmkdz\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.723508 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/88c9b59f-1809-4252-a8fc-cad965848dc0-proxy-tls\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.723533 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bec261f-fd1e-44f7-a402-cae34f722b6c-config\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.723557 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68dcdc32-485b-435c-81cc-43be463998bb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dnrhx\" (UID: \"68dcdc32-485b-435c-81cc-43be463998bb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.723963 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.724135 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb7b4d4-0441-43c4-9596-4d38b369d661-config\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.724167 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nm6v\" (UniqueName: \"kubernetes.io/projected/f5464a23-ec80-4717-bfe0-6efeab811853-kube-api-access-2nm6v\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.724203 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.724228 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/88c9b59f-1809-4252-a8fc-cad965848dc0-images\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.724250 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvhz9\" (UniqueName: \"kubernetes.io/projected/e9446a39-9776-4de8-9137-b8952d336419-kube-api-access-lvhz9\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.724276 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/288e6358-c74b-4597-8968-726a31365f82-encryption-config\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.724317 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bec261f-fd1e-44f7-a402-cae34f722b6c-config\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.724869 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb7b4d4-0441-43c4-9596-4d38b369d661-config\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.725255 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-auth-proxy-config\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.725290 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f87cb481-5261-4028-99e2-cd57ff6b61e1-config\") pod \"kube-apiserver-operator-766d6c64bb-gsfnm\" (UID: \"f87cb481-5261-4028-99e2-cd57ff6b61e1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.725315 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-client-ca\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.725353 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a67b9a8-ad8a-40e3-955c-53aed07a9140-config\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.725375 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726380 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726430 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9446a39-9776-4de8-9137-b8952d336419-metrics-certs\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726456 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb7b4d4-0441-43c4-9596-4d38b369d661-serving-cert\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726480 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-config\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726503 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b97ls\" (UniqueName: \"kubernetes.io/projected/9bec261f-fd1e-44f7-a402-cae34f722b6c-kube-api-access-b97ls\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726540 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca44490a-7fc8-478d-a6e2-670f49816b81-audit-dir\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726570 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-etcd-serving-ca\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726596 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca44490a-7fc8-478d-a6e2-670f49816b81-encryption-config\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726638 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/88c9b59f-1809-4252-a8fc-cad965848dc0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726666 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca44490a-7fc8-478d-a6e2-670f49816b81-serving-cert\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726706 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68dcdc32-485b-435c-81cc-43be463998bb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dnrhx\" (UID: \"68dcdc32-485b-435c-81cc-43be463998bb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726859 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9gf8\" (UniqueName: \"kubernetes.io/projected/9fd4241f-b523-4d66-bcdb-c3bb691765c9-kube-api-access-h9gf8\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726894 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsxpl\" (UniqueName: \"kubernetes.io/projected/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-kube-api-access-tsxpl\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.726923 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nfcq\" (UniqueName: \"kubernetes.io/projected/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-kube-api-access-9nfcq\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.727013 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/288e6358-c74b-4597-8968-726a31365f82-etcd-client\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.727038 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e9446a39-9776-4de8-9137-b8952d336419-default-certificate\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.727662 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.727793 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-audit\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.727945 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-auth-proxy-config\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.728046 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.728026 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca44490a-7fc8-478d-a6e2-670f49816b81-audit-dir\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.727037 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca44490a-7fc8-478d-a6e2-670f49816b81-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.728321 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7qdr\" (UniqueName: \"kubernetes.io/projected/5fb7b4d4-0441-43c4-9596-4d38b369d661-kube-api-access-x7qdr\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.728430 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f87cb481-5261-4028-99e2-cd57ff6b61e1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gsfnm\" (UID: \"f87cb481-5261-4028-99e2-cd57ff6b61e1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.728546 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-machine-approver-tls\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.728625 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5fb7b4d4-0441-43c4-9596-4d38b369d661-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.728676 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-oauth-serving-cert\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.728701 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c045a558-1990-432e-9965-e918f60aba14-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dprzc\" (UID: \"c045a558-1990-432e-9965-e918f60aba14\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.728736 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw8xf\" (UniqueName: \"kubernetes.io/projected/8a67b9a8-ad8a-40e3-955c-53aed07a9140-kube-api-access-cw8xf\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.728757 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729023 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-client-ca\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729056 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-trusted-ca-bundle\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729226 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f87cb481-5261-4028-99e2-cd57ff6b61e1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gsfnm\" (UID: \"f87cb481-5261-4028-99e2-cd57ff6b61e1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729484 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-config\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729538 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729626 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-trusted-ca-bundle\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729701 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5fb7b4d4-0441-43c4-9596-4d38b369d661-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729740 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5fb7b4d4-0441-43c4-9596-4d38b369d661-service-ca-bundle\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729795 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-config\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729858 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-config\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729886 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a67b9a8-ad8a-40e3-955c-53aed07a9140-serving-cert\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.729928 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a67b9a8-ad8a-40e3-955c-53aed07a9140-trusted-ca\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.730182 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.730258 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9bec261f-fd1e-44f7-a402-cae34f722b6c-images\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.730294 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-config\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.730371 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5fb7b4d4-0441-43c4-9596-4d38b369d661-service-ca-bundle\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.730384 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-config\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.731192 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9bec261f-fd1e-44f7-a402-cae34f722b6c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.731219 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9bec261f-fd1e-44f7-a402-cae34f722b6c-images\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.732495 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca44490a-7fc8-478d-a6e2-670f49816b81-serving-cert\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.732658 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.733108 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5464a23-ec80-4717-bfe0-6efeab811853-serving-cert\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.733122 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca44490a-7fc8-478d-a6e2-670f49816b81-etcd-client\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.734349 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.736839 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.745220 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb7b4d4-0441-43c4-9596-4d38b369d661-serving-cert\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.747430 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-f64ps"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.747633 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.748091 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.748438 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dwcjx"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.748534 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.748536 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.749314 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.749589 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.749718 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.749893 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.750524 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6dvvn"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.750784 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.750855 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-cfzn7"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.750942 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.751426 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.751813 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.752078 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.752316 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.752512 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.752596 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.752632 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-869fn"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.754232 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-jgw4c"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.755499 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.756089 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.757821 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.757864 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.758876 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8j9tz"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.760513 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mftbv"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.765424 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wws2s"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.767744 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.767621 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.769821 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.771234 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wmwcz"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.772816 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.773942 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.776353 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.777595 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jg4wj"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.778778 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.780827 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.782079 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.784427 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-f64ps"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.785610 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.787077 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.787275 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.788349 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-cqkbr"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.790332 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.791371 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rx6nv"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.793040 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.794034 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.795584 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-sjhjk"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.796650 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.798195 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.799829 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.803049 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dwcjx"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.804429 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.805653 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.807750 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.812446 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6dvvn"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.812508 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-l87tw"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.817083 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-l87tw"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.817172 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.818946 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.820065 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-kxvzk"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.820732 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.821772 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-s5xxd"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.822448 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s5xxd" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.823044 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kxvzk"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.824142 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s5xxd"] Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833288 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-trusted-ca-bundle\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833331 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f87cb481-5261-4028-99e2-cd57ff6b61e1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gsfnm\" (UID: \"f87cb481-5261-4028-99e2-cd57ff6b61e1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833357 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-trusted-ca-bundle\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833382 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-dir\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833425 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2xwq\" (UniqueName: \"kubernetes.io/projected/308c337e-1e28-4e34-9ccb-8ae546eee089-kube-api-access-j2xwq\") pod \"cluster-samples-operator-665b6dd947-qfcsk\" (UID: \"308c337e-1e28-4e34-9ccb-8ae546eee089\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833454 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4370babd-21ea-4c7e-81b9-cdd611127094-etcd-service-ca\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833477 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9bvp\" (UniqueName: \"kubernetes.io/projected/4370babd-21ea-4c7e-81b9-cdd611127094-kube-api-access-d9bvp\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833503 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kg7v\" (UniqueName: \"kubernetes.io/projected/16747491-2b7d-4cb1-841d-61c6f366cf8a-kube-api-access-2kg7v\") pod \"openshift-config-operator-7777fb866f-gv9x5\" (UID: \"16747491-2b7d-4cb1-841d-61c6f366cf8a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833529 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/308c337e-1e28-4e34-9ccb-8ae546eee089-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qfcsk\" (UID: \"308c337e-1e28-4e34-9ccb-8ae546eee089\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833554 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/288e6358-c74b-4597-8968-726a31365f82-audit-dir\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833578 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f745af09-93ed-4c94-8411-f14b0aaaf1cf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vsxx4\" (UID: \"f745af09-93ed-4c94-8411-f14b0aaaf1cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833606 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-policies\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833633 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4370babd-21ea-4c7e-81b9-cdd611127094-config\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833657 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmqf8\" (UniqueName: \"kubernetes.io/projected/8b5afeda-c180-4aa9-a831-a0840d495fa8-kube-api-access-hmqf8\") pod \"catalog-operator-68c6474976-dt8w6\" (UID: \"8b5afeda-c180-4aa9-a831-a0840d495fa8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833682 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-client-ca\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833707 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833728 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4370babd-21ea-4c7e-81b9-cdd611127094-etcd-client\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833577 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-dir\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833753 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a15050a-7cbd-40f6-a656-a68293c0878a-secret-volume\") pod \"collect-profiles-29484690-npncx\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.833778 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1795c220-db74-434e-9111-917ff6d95077-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-f64ps\" (UID: \"1795c220-db74-434e-9111-917ff6d95077\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834006 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2dad3db6-cddd-457d-8efa-908257ef7cc5-console-serving-cert\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834073 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9jtx\" (UniqueName: \"kubernetes.io/projected/c045a558-1990-432e-9965-e918f60aba14-kube-api-access-b9jtx\") pod \"machine-config-controller-84d6567774-dprzc\" (UID: \"c045a558-1990-432e-9965-e918f60aba14\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834142 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh722\" (UniqueName: \"kubernetes.io/projected/304c66b8-6187-47ac-9c57-235c634eaae4-kube-api-access-sh722\") pod \"migrator-59844c95c7-rwcrv\" (UID: \"304c66b8-6187-47ac-9c57-235c634eaae4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834182 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/288e6358-c74b-4597-8968-726a31365f82-node-pullsecrets\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834235 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rprpc\" (UniqueName: \"kubernetes.io/projected/2dad3db6-cddd-457d-8efa-908257ef7cc5-kube-api-access-rprpc\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834260 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj795\" (UniqueName: \"kubernetes.io/projected/3b9983ba-ed8d-4654-ba74-f25433aa7ee7-kube-api-access-gj795\") pod \"downloads-7954f5f757-jgw4c\" (UID: \"3b9983ba-ed8d-4654-ba74-f25433aa7ee7\") " pod="openshift-console/downloads-7954f5f757-jgw4c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834285 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-srv-cert\") pod \"olm-operator-6b444d44fb-b6c5m\" (UID: \"574225b6-a476-4591-9be9-ddd94e5281ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834342 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-profile-collector-cert\") pod \"catalog-operator-68c6474976-dt8w6\" (UID: \"8b5afeda-c180-4aa9-a831-a0840d495fa8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834368 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834391 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834429 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/288e6358-c74b-4597-8968-726a31365f82-serving-cert\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834452 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf6nt\" (UniqueName: \"kubernetes.io/projected/213d34f5-75cd-459c-9e56-2938fe5e3950-kube-api-access-lf6nt\") pod \"marketplace-operator-79b997595-6dvvn\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834495 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/288e6358-c74b-4597-8968-726a31365f82-node-pullsecrets\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834823 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-trusted-ca-bundle\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834910 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-policies\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.834968 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c045a558-1990-432e-9965-e918f60aba14-proxy-tls\") pod \"machine-config-controller-84d6567774-dprzc\" (UID: \"c045a558-1990-432e-9965-e918f60aba14\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.835007 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65f71c2e-ab34-4d33-905f-609555dab78c-serving-cert\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.835034 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dm8p\" (UniqueName: \"kubernetes.io/projected/88c9b59f-1809-4252-a8fc-cad965848dc0-kube-api-access-8dm8p\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.835073 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d80367da-a8d4-4b76-9b33-5526aad85229-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s4fb2\" (UID: \"d80367da-a8d4-4b76-9b33-5526aad85229\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.835092 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krt27\" (UniqueName: \"kubernetes.io/projected/1795c220-db74-434e-9111-917ff6d95077-kube-api-access-krt27\") pod \"multus-admission-controller-857f4d67dd-f64ps\" (UID: \"1795c220-db74-434e-9111-917ff6d95077\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.835123 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.835142 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.835139 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/288e6358-c74b-4597-8968-726a31365f82-audit-dir\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.835306 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmkdz\" (UniqueName: \"kubernetes.io/projected/288e6358-c74b-4597-8968-726a31365f82-kube-api-access-qmkdz\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.835363 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/88c9b59f-1809-4252-a8fc-cad965848dc0-proxy-tls\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.835819 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.836097 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/88c9b59f-1809-4252-a8fc-cad965848dc0-images\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.836153 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvhz9\" (UniqueName: \"kubernetes.io/projected/e9446a39-9776-4de8-9137-b8952d336419-kube-api-access-lvhz9\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.836335 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.836440 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/288e6358-c74b-4597-8968-726a31365f82-encryption-config\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.836485 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.836526 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9446a39-9776-4de8-9137-b8952d336419-metrics-certs\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.836646 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjztm\" (UniqueName: \"kubernetes.io/projected/0a15050a-7cbd-40f6-a656-a68293c0878a-kube-api-access-fjztm\") pod \"collect-profiles-29484690-npncx\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.836712 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-etcd-serving-ca\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.836762 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/88c9b59f-1809-4252-a8fc-cad965848dc0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.836807 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9gf8\" (UniqueName: \"kubernetes.io/projected/9fd4241f-b523-4d66-bcdb-c3bb691765c9-kube-api-access-h9gf8\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.836844 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/88c9b59f-1809-4252-a8fc-cad965848dc0-images\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.837001 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-trusted-ca-bundle\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.837093 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a15050a-7cbd-40f6-a656-a68293c0878a-config-volume\") pod \"collect-profiles-29484690-npncx\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.837145 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d80367da-a8d4-4b76-9b33-5526aad85229-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s4fb2\" (UID: \"d80367da-a8d4-4b76-9b33-5526aad85229\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.837185 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-audit\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.837347 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-serving-cert\") pod \"service-ca-operator-777779d784-6zpzc\" (UID: \"51b2c452-70bd-4a3e-bd74-c7cd6513fd45\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.837368 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-profile-collector-cert\") pod \"olm-operator-6b444d44fb-b6c5m\" (UID: \"574225b6-a476-4591-9be9-ddd94e5281ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.837412 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/88c9b59f-1809-4252-a8fc-cad965848dc0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.837389 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-srv-cert\") pod \"catalog-operator-68c6474976-dt8w6\" (UID: \"8b5afeda-c180-4aa9-a831-a0840d495fa8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.837770 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-etcd-serving-ca\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.838004 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-audit\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.838071 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-oauth-serving-cert\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.838447 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c045a558-1990-432e-9965-e918f60aba14-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dprzc\" (UID: \"c045a558-1990-432e-9965-e918f60aba14\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.838597 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6dvvn\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.838826 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-config\") pod \"service-ca-operator-777779d784-6zpzc\" (UID: \"51b2c452-70bd-4a3e-bd74-c7cd6513fd45\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.838931 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3d1224e-a43e-4dca-8ad0-239dc50b6d58-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fmhx8\" (UID: \"a3d1224e-a43e-4dca-8ad0-239dc50b6d58\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839039 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d7ccd8e-549f-4b23-bcec-4a1c14e10478-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vqrw2\" (UID: \"8d7ccd8e-549f-4b23-bcec-4a1c14e10478\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839144 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx7xc\" (UniqueName: \"kubernetes.io/projected/f745af09-93ed-4c94-8411-f14b0aaaf1cf-kube-api-access-lx7xc\") pod \"openshift-controller-manager-operator-756b6f6bc6-vsxx4\" (UID: \"f745af09-93ed-4c94-8411-f14b0aaaf1cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839247 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2dad3db6-cddd-457d-8efa-908257ef7cc5-console-serving-cert\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839175 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-oauth-serving-cert\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.838897 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/288e6358-c74b-4597-8968-726a31365f82-serving-cert\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839490 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839573 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-config\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839657 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a67b9a8-ad8a-40e3-955c-53aed07a9140-serving-cert\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839715 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839791 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a67b9a8-ad8a-40e3-955c-53aed07a9140-trusted-ca\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839869 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4370babd-21ea-4c7e-81b9-cdd611127094-etcd-ca\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839944 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-config\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840017 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d80367da-a8d4-4b76-9b33-5526aad85229-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s4fb2\" (UID: \"d80367da-a8d4-4b76-9b33-5526aad85229\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840096 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7t6g\" (UniqueName: \"kubernetes.io/projected/a3d1224e-a43e-4dca-8ad0-239dc50b6d58-kube-api-access-p7t6g\") pod \"kube-storage-version-migrator-operator-b67b599dd-fmhx8\" (UID: \"a3d1224e-a43e-4dca-8ad0-239dc50b6d58\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840191 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-service-ca\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.839981 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/308c337e-1e28-4e34-9ccb-8ae546eee089-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qfcsk\" (UID: \"308c337e-1e28-4e34-9ccb-8ae546eee089\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840265 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9hxd\" (UniqueName: \"kubernetes.io/projected/68dcdc32-485b-435c-81cc-43be463998bb-kube-api-access-c9hxd\") pod \"openshift-apiserver-operator-796bbdcf4f-dnrhx\" (UID: \"68dcdc32-485b-435c-81cc-43be463998bb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840332 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-console-config\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840356 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/16747491-2b7d-4cb1-841d-61c6f366cf8a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gv9x5\" (UID: \"16747491-2b7d-4cb1-841d-61c6f366cf8a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840376 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840416 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtv7h\" (UniqueName: \"kubernetes.io/projected/574225b6-a476-4591-9be9-ddd94e5281ef-kube-api-access-jtv7h\") pod \"olm-operator-6b444d44fb-b6c5m\" (UID: \"574225b6-a476-4591-9be9-ddd94e5281ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840439 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9446a39-9776-4de8-9137-b8952d336419-service-ca-bundle\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840458 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840479 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htst2\" (UniqueName: \"kubernetes.io/projected/4f61e177-ea8e-4f61-8ba2-67906f08a00c-kube-api-access-htst2\") pod \"dns-operator-744455d44c-rx6nv\" (UID: \"4f61e177-ea8e-4f61-8ba2-67906f08a00c\") " pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840509 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m29d2\" (UniqueName: \"kubernetes.io/projected/65f71c2e-ab34-4d33-905f-609555dab78c-kube-api-access-m29d2\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840526 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-tmpfs\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840538 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-config\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840550 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2dad3db6-cddd-457d-8efa-908257ef7cc5-console-oauth-config\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840590 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4f61e177-ea8e-4f61-8ba2-67906f08a00c-metrics-tls\") pod \"dns-operator-744455d44c-rx6nv\" (UID: \"4f61e177-ea8e-4f61-8ba2-67906f08a00c\") " pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840606 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-apiservice-cert\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840625 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-image-import-ca\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840642 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16747491-2b7d-4cb1-841d-61c6f366cf8a-serving-cert\") pod \"openshift-config-operator-7777fb866f-gv9x5\" (UID: \"16747491-2b7d-4cb1-841d-61c6f366cf8a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840658 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e9446a39-9776-4de8-9137-b8952d336419-stats-auth\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840674 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f745af09-93ed-4c94-8411-f14b0aaaf1cf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vsxx4\" (UID: \"f745af09-93ed-4c94-8411-f14b0aaaf1cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840714 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-certs\") pod \"machine-config-server-cfzn7\" (UID: \"061a736b-3663-4a73-9836-19724fc4bb71\") " pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840742 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68dcdc32-485b-435c-81cc-43be463998bb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dnrhx\" (UID: \"68dcdc32-485b-435c-81cc-43be463998bb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840760 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840780 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4370babd-21ea-4c7e-81b9-cdd611127094-serving-cert\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840800 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f87cb481-5261-4028-99e2-cd57ff6b61e1-config\") pod \"kube-apiserver-operator-766d6c64bb-gsfnm\" (UID: \"f87cb481-5261-4028-99e2-cd57ff6b61e1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840815 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-node-bootstrap-token\") pod \"machine-config-server-cfzn7\" (UID: \"061a736b-3663-4a73-9836-19724fc4bb71\") " pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840847 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a67b9a8-ad8a-40e3-955c-53aed07a9140-config\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840877 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840901 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-webhook-cert\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840928 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdhbq\" (UniqueName: \"kubernetes.io/projected/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-kube-api-access-sdhbq\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840709 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a67b9a8-ad8a-40e3-955c-53aed07a9140-trusted-ca\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840951 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6dvvn\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.840987 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68dcdc32-485b-435c-81cc-43be463998bb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dnrhx\" (UID: \"68dcdc32-485b-435c-81cc-43be463998bb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841023 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/288e6358-c74b-4597-8968-726a31365f82-etcd-client\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841043 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e9446a39-9776-4de8-9137-b8952d336419-default-certificate\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841065 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d1224e-a43e-4dca-8ad0-239dc50b6d58-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fmhx8\" (UID: \"a3d1224e-a43e-4dca-8ad0-239dc50b6d58\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841094 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f87cb481-5261-4028-99e2-cd57ff6b61e1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gsfnm\" (UID: \"f87cb481-5261-4028-99e2-cd57ff6b61e1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841114 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw8xf\" (UniqueName: \"kubernetes.io/projected/8a67b9a8-ad8a-40e3-955c-53aed07a9140-kube-api-access-cw8xf\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841133 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841154 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqmnt\" (UniqueName: \"kubernetes.io/projected/8d7ccd8e-549f-4b23-bcec-4a1c14e10478-kube-api-access-lqmnt\") pod \"package-server-manager-789f6589d5-vqrw2\" (UID: \"8d7ccd8e-549f-4b23-bcec-4a1c14e10478\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841168 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/16747491-2b7d-4cb1-841d-61c6f366cf8a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gv9x5\" (UID: \"16747491-2b7d-4cb1-841d-61c6f366cf8a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841173 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b42gj\" (UniqueName: \"kubernetes.io/projected/061a736b-3663-4a73-9836-19724fc4bb71-kube-api-access-b42gj\") pod \"machine-config-server-cfzn7\" (UID: \"061a736b-3663-4a73-9836-19724fc4bb71\") " pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841250 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-console-config\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841533 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68dcdc32-485b-435c-81cc-43be463998bb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dnrhx\" (UID: \"68dcdc32-485b-435c-81cc-43be463998bb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841727 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2dad3db6-cddd-457d-8efa-908257ef7cc5-service-ca\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.841945 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a67b9a8-ad8a-40e3-955c-53aed07a9140-config\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.842429 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhgtt\" (UniqueName: \"kubernetes.io/projected/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-kube-api-access-bhgtt\") pod \"service-ca-operator-777779d784-6zpzc\" (UID: \"51b2c452-70bd-4a3e-bd74-c7cd6513fd45\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.842507 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9446a39-9776-4de8-9137-b8952d336419-service-ca-bundle\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.842952 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/288e6358-c74b-4597-8968-726a31365f82-image-import-ca\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.842966 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/88c9b59f-1809-4252-a8fc-cad965848dc0-proxy-tls\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.843135 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.843167 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9446a39-9776-4de8-9137-b8952d336419-metrics-certs\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.843362 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.843713 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.843888 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2dad3db6-cddd-457d-8efa-908257ef7cc5-console-oauth-config\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.845100 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e9446a39-9776-4de8-9137-b8952d336419-default-certificate\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.845256 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.845334 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c045a558-1990-432e-9965-e918f60aba14-proxy-tls\") pod \"machine-config-controller-84d6567774-dprzc\" (UID: \"c045a558-1990-432e-9965-e918f60aba14\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.845478 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c045a558-1990-432e-9965-e918f60aba14-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dprzc\" (UID: \"c045a558-1990-432e-9965-e918f60aba14\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.845561 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/288e6358-c74b-4597-8968-726a31365f82-encryption-config\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.845652 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16747491-2b7d-4cb1-841d-61c6f366cf8a-serving-cert\") pod \"openshift-config-operator-7777fb866f-gv9x5\" (UID: \"16747491-2b7d-4cb1-841d-61c6f366cf8a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.845666 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/288e6358-c74b-4597-8968-726a31365f82-etcd-client\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.846194 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a67b9a8-ad8a-40e3-955c-53aed07a9140-serving-cert\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.846311 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.846740 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.846769 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68dcdc32-485b-435c-81cc-43be463998bb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dnrhx\" (UID: \"68dcdc32-485b-435c-81cc-43be463998bb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.847062 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e9446a39-9776-4de8-9137-b8952d336419-stats-auth\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.847201 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.847256 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.866970 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.888604 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.897985 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65f71c2e-ab34-4d33-905f-609555dab78c-serving-cert\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.907362 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.910689 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-config\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.927479 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.935964 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-client-ca\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.943647 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-node-bootstrap-token\") pod \"machine-config-server-cfzn7\" (UID: \"061a736b-3663-4a73-9836-19724fc4bb71\") " pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.943701 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-webhook-cert\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.943743 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdhbq\" (UniqueName: \"kubernetes.io/projected/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-kube-api-access-sdhbq\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.943777 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6dvvn\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.943834 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d1224e-a43e-4dca-8ad0-239dc50b6d58-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fmhx8\" (UID: \"a3d1224e-a43e-4dca-8ad0-239dc50b6d58\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.943900 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqmnt\" (UniqueName: \"kubernetes.io/projected/8d7ccd8e-549f-4b23-bcec-4a1c14e10478-kube-api-access-lqmnt\") pod \"package-server-manager-789f6589d5-vqrw2\" (UID: \"8d7ccd8e-549f-4b23-bcec-4a1c14e10478\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.943931 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b42gj\" (UniqueName: \"kubernetes.io/projected/061a736b-3663-4a73-9836-19724fc4bb71-kube-api-access-b42gj\") pod \"machine-config-server-cfzn7\" (UID: \"061a736b-3663-4a73-9836-19724fc4bb71\") " pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.943966 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhgtt\" (UniqueName: \"kubernetes.io/projected/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-kube-api-access-bhgtt\") pod \"service-ca-operator-777779d784-6zpzc\" (UID: \"51b2c452-70bd-4a3e-bd74-c7cd6513fd45\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944029 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4370babd-21ea-4c7e-81b9-cdd611127094-etcd-service-ca\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944078 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9bvp\" (UniqueName: \"kubernetes.io/projected/4370babd-21ea-4c7e-81b9-cdd611127094-kube-api-access-d9bvp\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944125 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f745af09-93ed-4c94-8411-f14b0aaaf1cf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vsxx4\" (UID: \"f745af09-93ed-4c94-8411-f14b0aaaf1cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944165 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4370babd-21ea-4c7e-81b9-cdd611127094-config\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944232 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmqf8\" (UniqueName: \"kubernetes.io/projected/8b5afeda-c180-4aa9-a831-a0840d495fa8-kube-api-access-hmqf8\") pod \"catalog-operator-68c6474976-dt8w6\" (UID: \"8b5afeda-c180-4aa9-a831-a0840d495fa8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944273 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4370babd-21ea-4c7e-81b9-cdd611127094-etcd-client\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944308 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a15050a-7cbd-40f6-a656-a68293c0878a-secret-volume\") pod \"collect-profiles-29484690-npncx\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944347 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1795c220-db74-434e-9111-917ff6d95077-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-f64ps\" (UID: \"1795c220-db74-434e-9111-917ff6d95077\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944468 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-srv-cert\") pod \"olm-operator-6b444d44fb-b6c5m\" (UID: \"574225b6-a476-4591-9be9-ddd94e5281ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944508 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-profile-collector-cert\") pod \"catalog-operator-68c6474976-dt8w6\" (UID: \"8b5afeda-c180-4aa9-a831-a0840d495fa8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944538 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf6nt\" (UniqueName: \"kubernetes.io/projected/213d34f5-75cd-459c-9e56-2938fe5e3950-kube-api-access-lf6nt\") pod \"marketplace-operator-79b997595-6dvvn\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944586 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d80367da-a8d4-4b76-9b33-5526aad85229-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s4fb2\" (UID: \"d80367da-a8d4-4b76-9b33-5526aad85229\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944624 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krt27\" (UniqueName: \"kubernetes.io/projected/1795c220-db74-434e-9111-917ff6d95077-kube-api-access-krt27\") pod \"multus-admission-controller-857f4d67dd-f64ps\" (UID: \"1795c220-db74-434e-9111-917ff6d95077\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944674 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjztm\" (UniqueName: \"kubernetes.io/projected/0a15050a-7cbd-40f6-a656-a68293c0878a-kube-api-access-fjztm\") pod \"collect-profiles-29484690-npncx\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944713 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a15050a-7cbd-40f6-a656-a68293c0878a-config-volume\") pod \"collect-profiles-29484690-npncx\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944753 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d80367da-a8d4-4b76-9b33-5526aad85229-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s4fb2\" (UID: \"d80367da-a8d4-4b76-9b33-5526aad85229\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944775 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-serving-cert\") pod \"service-ca-operator-777779d784-6zpzc\" (UID: \"51b2c452-70bd-4a3e-bd74-c7cd6513fd45\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944795 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-profile-collector-cert\") pod \"olm-operator-6b444d44fb-b6c5m\" (UID: \"574225b6-a476-4591-9be9-ddd94e5281ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944814 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-srv-cert\") pod \"catalog-operator-68c6474976-dt8w6\" (UID: \"8b5afeda-c180-4aa9-a831-a0840d495fa8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944844 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6dvvn\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944875 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-config\") pod \"service-ca-operator-777779d784-6zpzc\" (UID: \"51b2c452-70bd-4a3e-bd74-c7cd6513fd45\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944906 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3d1224e-a43e-4dca-8ad0-239dc50b6d58-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fmhx8\" (UID: \"a3d1224e-a43e-4dca-8ad0-239dc50b6d58\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944931 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d7ccd8e-549f-4b23-bcec-4a1c14e10478-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vqrw2\" (UID: \"8d7ccd8e-549f-4b23-bcec-4a1c14e10478\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944957 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx7xc\" (UniqueName: \"kubernetes.io/projected/f745af09-93ed-4c94-8411-f14b0aaaf1cf-kube-api-access-lx7xc\") pod \"openshift-controller-manager-operator-756b6f6bc6-vsxx4\" (UID: \"f745af09-93ed-4c94-8411-f14b0aaaf1cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.944980 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4370babd-21ea-4c7e-81b9-cdd611127094-etcd-ca\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.945003 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d80367da-a8d4-4b76-9b33-5526aad85229-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s4fb2\" (UID: \"d80367da-a8d4-4b76-9b33-5526aad85229\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.945029 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7t6g\" (UniqueName: \"kubernetes.io/projected/a3d1224e-a43e-4dca-8ad0-239dc50b6d58-kube-api-access-p7t6g\") pod \"kube-storage-version-migrator-operator-b67b599dd-fmhx8\" (UID: \"a3d1224e-a43e-4dca-8ad0-239dc50b6d58\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.945084 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtv7h\" (UniqueName: \"kubernetes.io/projected/574225b6-a476-4591-9be9-ddd94e5281ef-kube-api-access-jtv7h\") pod \"olm-operator-6b444d44fb-b6c5m\" (UID: \"574225b6-a476-4591-9be9-ddd94e5281ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.945127 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htst2\" (UniqueName: \"kubernetes.io/projected/4f61e177-ea8e-4f61-8ba2-67906f08a00c-kube-api-access-htst2\") pod \"dns-operator-744455d44c-rx6nv\" (UID: \"4f61e177-ea8e-4f61-8ba2-67906f08a00c\") " pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.945169 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-tmpfs\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.945199 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4f61e177-ea8e-4f61-8ba2-67906f08a00c-metrics-tls\") pod \"dns-operator-744455d44c-rx6nv\" (UID: \"4f61e177-ea8e-4f61-8ba2-67906f08a00c\") " pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.945219 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-apiservice-cert\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.945244 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f745af09-93ed-4c94-8411-f14b0aaaf1cf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vsxx4\" (UID: \"f745af09-93ed-4c94-8411-f14b0aaaf1cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.945267 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-certs\") pod \"machine-config-server-cfzn7\" (UID: \"061a736b-3663-4a73-9836-19724fc4bb71\") " pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.945321 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4370babd-21ea-4c7e-81b9-cdd611127094-serving-cert\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.946586 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-tmpfs\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.947958 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.967364 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 22 11:42:40 crc kubenswrapper[4874]: I0122 11:42:40.986696 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.007877 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.017373 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f87cb481-5261-4028-99e2-cd57ff6b61e1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gsfnm\" (UID: \"f87cb481-5261-4028-99e2-cd57ff6b61e1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.027510 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.034284 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f87cb481-5261-4028-99e2-cd57ff6b61e1-config\") pod \"kube-apiserver-operator-766d6c64bb-gsfnm\" (UID: \"f87cb481-5261-4028-99e2-cd57ff6b61e1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.047326 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.058079 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3d1224e-a43e-4dca-8ad0-239dc50b6d58-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fmhx8\" (UID: \"a3d1224e-a43e-4dca-8ad0-239dc50b6d58\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.068138 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.086900 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.107586 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.115948 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d1224e-a43e-4dca-8ad0-239dc50b6d58-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fmhx8\" (UID: \"a3d1224e-a43e-4dca-8ad0-239dc50b6d58\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.126881 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.148145 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.168122 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.187917 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.208035 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.226958 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.247595 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.268141 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.278169 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4f61e177-ea8e-4f61-8ba2-67906f08a00c-metrics-tls\") pod \"dns-operator-744455d44c-rx6nv\" (UID: \"4f61e177-ea8e-4f61-8ba2-67906f08a00c\") " pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.288047 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.307866 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.328205 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.347603 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.367665 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.387385 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.395443 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d80367da-a8d4-4b76-9b33-5526aad85229-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s4fb2\" (UID: \"d80367da-a8d4-4b76-9b33-5526aad85229\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.407472 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.419309 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d80367da-a8d4-4b76-9b33-5526aad85229-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s4fb2\" (UID: \"d80367da-a8d4-4b76-9b33-5526aad85229\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.427871 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.447709 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.467653 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.487551 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.508501 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.527765 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.557788 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.568566 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.588150 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.607698 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.628123 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.640994 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f745af09-93ed-4c94-8411-f14b0aaaf1cf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vsxx4\" (UID: \"f745af09-93ed-4c94-8411-f14b0aaaf1cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.648224 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.655588 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f745af09-93ed-4c94-8411-f14b0aaaf1cf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vsxx4\" (UID: \"f745af09-93ed-4c94-8411-f14b0aaaf1cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.667798 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.711803 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nm6v\" (UniqueName: \"kubernetes.io/projected/f5464a23-ec80-4717-bfe0-6efeab811853-kube-api-access-2nm6v\") pod \"controller-manager-879f6c89f-swcnq\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.732660 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkqws\" (UniqueName: \"kubernetes.io/projected/ca44490a-7fc8-478d-a6e2-670f49816b81-kube-api-access-vkqws\") pod \"apiserver-7bbb656c7d-fl8nd\" (UID: \"ca44490a-7fc8-478d-a6e2-670f49816b81\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.745692 4874 request.go:700] Waited for 1.017661557s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/serviceaccounts/machine-api-operator/token Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.749469 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nfcq\" (UniqueName: \"kubernetes.io/projected/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-kube-api-access-9nfcq\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.762382 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.767257 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b97ls\" (UniqueName: \"kubernetes.io/projected/9bec261f-fd1e-44f7-a402-cae34f722b6c-kube-api-access-b97ls\") pod \"machine-api-operator-5694c8668f-wmwcz\" (UID: \"9bec261f-fd1e-44f7-a402-cae34f722b6c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.781596 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsxpl\" (UniqueName: \"kubernetes.io/projected/6c9561fa-20b6-4f87-aacc-cf0e0665ffa4-kube-api-access-tsxpl\") pod \"machine-approver-56656f9798-ssbbs\" (UID: \"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.804239 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0d49cfd6-9c1b-481f-a9d3-07a99661cf9d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-7h7c7\" (UID: \"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.808125 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.814010 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.827749 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.830887 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.849779 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.851851 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4370babd-21ea-4c7e-81b9-cdd611127094-etcd-ca\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.861123 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4370babd-21ea-4c7e-81b9-cdd611127094-serving-cert\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.862674 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.867797 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.879300 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4370babd-21ea-4c7e-81b9-cdd611127094-etcd-client\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.887226 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.890440 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.895126 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4370babd-21ea-4c7e-81b9-cdd611127094-config\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:41 crc kubenswrapper[4874]: W0122 11:42:41.903889 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c9561fa_20b6_4f87_aacc_cf0e0665ffa4.slice/crio-c9acde3f2bdd745a092777b2d7badcee478c91b536adde065c962ce05336fe43 WatchSource:0}: Error finding container c9acde3f2bdd745a092777b2d7badcee478c91b536adde065c962ce05336fe43: Status 404 returned error can't find the container with id c9acde3f2bdd745a092777b2d7badcee478c91b536adde065c962ce05336fe43 Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.911050 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.915007 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4370babd-21ea-4c7e-81b9-cdd611127094-etcd-service-ca\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.930676 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.944418 4874 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.944515 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-node-bootstrap-token podName:061a736b-3663-4a73-9836-19724fc4bb71 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.444491104 +0000 UTC m=+136.289562174 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-node-bootstrap-token") pod "machine-config-server-cfzn7" (UID: "061a736b-3663-4a73-9836-19724fc4bb71") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.945064 4874 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.945155 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d7ccd8e-549f-4b23-bcec-4a1c14e10478-package-server-manager-serving-cert podName:8d7ccd8e-549f-4b23-bcec-4a1c14e10478 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.445127794 +0000 UTC m=+136.290198904 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8d7ccd8e-549f-4b23-bcec-4a1c14e10478-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-vqrw2" (UID: "8d7ccd8e-549f-4b23-bcec-4a1c14e10478") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.945183 4874 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.945216 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1795c220-db74-434e-9111-917ff6d95077-webhook-certs podName:1795c220-db74-434e-9111-917ff6d95077 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.445207567 +0000 UTC m=+136.290278637 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/1795c220-db74-434e-9111-917ff6d95077-webhook-certs") pod "multus-admission-controller-857f4d67dd-f64ps" (UID: "1795c220-db74-434e-9111-917ff6d95077") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.946353 4874 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.946419 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-apiservice-cert podName:1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.446406814 +0000 UTC m=+136.291477884 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-apiservice-cert") pod "packageserver-d55dfcdfc-6pr5c" (UID: "1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.947636 4874 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.947724 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-certs podName:061a736b-3663-4a73-9836-19724fc4bb71 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.447701455 +0000 UTC m=+136.292772565 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-certs") pod "machine-config-server-cfzn7" (UID: "061a736b-3663-4a73-9836-19724fc4bb71") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.950638 4874 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.950681 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-serving-cert podName:51b2c452-70bd-4a3e-bd74-c7cd6513fd45 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.450671869 +0000 UTC m=+136.295742939 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-serving-cert") pod "service-ca-operator-777779d784-6zpzc" (UID: "51b2c452-70bd-4a3e-bd74-c7cd6513fd45") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951178 4874 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951199 4874 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951227 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-profile-collector-cert podName:574225b6-a476-4591-9be9-ddd94e5281ef nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.451216496 +0000 UTC m=+136.296287616 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-profile-collector-cert") pod "olm-operator-6b444d44fb-b6c5m" (UID: "574225b6-a476-4591-9be9-ddd94e5281ef") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951252 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-srv-cert podName:574225b6-a476-4591-9be9-ddd94e5281ef nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.451240787 +0000 UTC m=+136.296311857 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-srv-cert") pod "olm-operator-6b444d44fb-b6c5m" (UID: "574225b6-a476-4591-9be9-ddd94e5281ef") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951264 4874 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951270 4874 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951269 4874 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951287 4874 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951300 4874 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951310 4874 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951300 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-trusted-ca podName:213d34f5-75cd-459c-9e56-2938fe5e3950 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.451290959 +0000 UTC m=+136.296362099 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-trusted-ca") pod "marketplace-operator-79b997595-6dvvn" (UID: "213d34f5-75cd-459c-9e56-2938fe5e3950") : failed to sync configmap cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951328 4874 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951328 4874 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951363 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-srv-cert podName:8b5afeda-c180-4aa9-a831-a0840d495fa8 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.45133641 +0000 UTC m=+136.296407480 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-srv-cert") pod "catalog-operator-68c6474976-dt8w6" (UID: "8b5afeda-c180-4aa9-a831-a0840d495fa8") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951380 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-webhook-cert podName:1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.45137343 +0000 UTC m=+136.296444490 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-webhook-cert") pod "packageserver-d55dfcdfc-6pr5c" (UID: "1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951421 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-config podName:51b2c452-70bd-4a3e-bd74-c7cd6513fd45 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.451387001 +0000 UTC m=+136.296458071 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-config") pod "service-ca-operator-777779d784-6zpzc" (UID: "51b2c452-70bd-4a3e-bd74-c7cd6513fd45") : failed to sync configmap cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951439 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a15050a-7cbd-40f6-a656-a68293c0878a-secret-volume podName:0a15050a-7cbd-40f6-a656-a68293c0878a nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.451431642 +0000 UTC m=+136.296502712 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-volume" (UniqueName: "kubernetes.io/secret/0a15050a-7cbd-40f6-a656-a68293c0878a-secret-volume") pod "collect-profiles-29484690-npncx" (UID: "0a15050a-7cbd-40f6-a656-a68293c0878a") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951451 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-operator-metrics podName:213d34f5-75cd-459c-9e56-2938fe5e3950 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.451445152 +0000 UTC m=+136.296516222 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-operator-metrics") pod "marketplace-operator-79b997595-6dvvn" (UID: "213d34f5-75cd-459c-9e56-2938fe5e3950") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951464 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-profile-collector-cert podName:8b5afeda-c180-4aa9-a831-a0840d495fa8 nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.451458353 +0000 UTC m=+136.296529423 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-profile-collector-cert") pod "catalog-operator-68c6474976-dt8w6" (UID: "8b5afeda-c180-4aa9-a831-a0840d495fa8") : failed to sync secret cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: E0122 11:42:41.951476 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0a15050a-7cbd-40f6-a656-a68293c0878a-config-volume podName:0a15050a-7cbd-40f6-a656-a68293c0878a nodeName:}" failed. No retries permitted until 2026-01-22 11:42:42.451470603 +0000 UTC m=+136.296541673 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/0a15050a-7cbd-40f6-a656-a68293c0878a-config-volume") pod "collect-profiles-29484690-npncx" (UID: "0a15050a-7cbd-40f6-a656-a68293c0878a") : failed to sync configmap cache: timed out waiting for the condition Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.952999 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.982282 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-swcnq"] Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.986147 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7qdr\" (UniqueName: \"kubernetes.io/projected/5fb7b4d4-0441-43c4-9596-4d38b369d661-kube-api-access-x7qdr\") pod \"authentication-operator-69f744f599-869fn\" (UID: \"5fb7b4d4-0441-43c4-9596-4d38b369d661\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:41 crc kubenswrapper[4874]: I0122 11:42:41.987986 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 22 11:42:41 crc kubenswrapper[4874]: W0122 11:42:41.994435 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5464a23_ec80_4717_bfe0_6efeab811853.slice/crio-d437fa062da853d4db086da0d6818e76b56a3a9b163a7e48bb221915ade96982 WatchSource:0}: Error finding container d437fa062da853d4db086da0d6818e76b56a3a9b163a7e48bb221915ade96982: Status 404 returned error can't find the container with id d437fa062da853d4db086da0d6818e76b56a3a9b163a7e48bb221915ade96982 Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.007726 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.011205 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7"] Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.027222 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.048987 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.067599 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.087388 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.108946 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.128014 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.144796 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.146769 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.168583 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.187925 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.207499 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.228001 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.248319 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.254798 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd"] Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.262625 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wmwcz"] Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.266996 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.287425 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.308661 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-869fn"] Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.309267 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.327185 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.354642 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.368475 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.387900 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.408665 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.427427 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.448985 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.469763 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.470256 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a15050a-7cbd-40f6-a656-a68293c0878a-secret-volume\") pod \"collect-profiles-29484690-npncx\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.470353 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1795c220-db74-434e-9111-917ff6d95077-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-f64ps\" (UID: \"1795c220-db74-434e-9111-917ff6d95077\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.470513 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-srv-cert\") pod \"olm-operator-6b444d44fb-b6c5m\" (UID: \"574225b6-a476-4591-9be9-ddd94e5281ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.470596 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-profile-collector-cert\") pod \"catalog-operator-68c6474976-dt8w6\" (UID: \"8b5afeda-c180-4aa9-a831-a0840d495fa8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.470879 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a15050a-7cbd-40f6-a656-a68293c0878a-config-volume\") pod \"collect-profiles-29484690-npncx\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.472516 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a15050a-7cbd-40f6-a656-a68293c0878a-config-volume\") pod \"collect-profiles-29484690-npncx\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.472581 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-serving-cert\") pod \"service-ca-operator-777779d784-6zpzc\" (UID: \"51b2c452-70bd-4a3e-bd74-c7cd6513fd45\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.472623 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-profile-collector-cert\") pod \"olm-operator-6b444d44fb-b6c5m\" (UID: \"574225b6-a476-4591-9be9-ddd94e5281ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.472646 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-srv-cert\") pod \"catalog-operator-68c6474976-dt8w6\" (UID: \"8b5afeda-c180-4aa9-a831-a0840d495fa8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.472685 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6dvvn\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.472712 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-config\") pod \"service-ca-operator-777779d784-6zpzc\" (UID: \"51b2c452-70bd-4a3e-bd74-c7cd6513fd45\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.472738 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d7ccd8e-549f-4b23-bcec-4a1c14e10478-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vqrw2\" (UID: \"8d7ccd8e-549f-4b23-bcec-4a1c14e10478\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.472928 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-apiservice-cert\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.472992 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-certs\") pod \"machine-config-server-cfzn7\" (UID: \"061a736b-3663-4a73-9836-19724fc4bb71\") " pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.473492 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-node-bootstrap-token\") pod \"machine-config-server-cfzn7\" (UID: \"061a736b-3663-4a73-9836-19724fc4bb71\") " pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.473528 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-webhook-cert\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.473565 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6dvvn\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.473808 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6dvvn\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.476008 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a15050a-7cbd-40f6-a656-a68293c0878a-secret-volume\") pod \"collect-profiles-29484690-npncx\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.476105 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-srv-cert\") pod \"olm-operator-6b444d44fb-b6c5m\" (UID: \"574225b6-a476-4591-9be9-ddd94e5281ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.476224 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-profile-collector-cert\") pod \"catalog-operator-68c6474976-dt8w6\" (UID: \"8b5afeda-c180-4aa9-a831-a0840d495fa8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.476246 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1795c220-db74-434e-9111-917ff6d95077-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-f64ps\" (UID: \"1795c220-db74-434e-9111-917ff6d95077\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.476830 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-webhook-cert\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.477902 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/574225b6-a476-4591-9be9-ddd94e5281ef-profile-collector-cert\") pod \"olm-operator-6b444d44fb-b6c5m\" (UID: \"574225b6-a476-4591-9be9-ddd94e5281ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.478026 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6dvvn\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.478440 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d7ccd8e-549f-4b23-bcec-4a1c14e10478-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vqrw2\" (UID: \"8d7ccd8e-549f-4b23-bcec-4a1c14e10478\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.478632 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-apiservice-cert\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.478773 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8b5afeda-c180-4aa9-a831-a0840d495fa8-srv-cert\") pod \"catalog-operator-68c6474976-dt8w6\" (UID: \"8b5afeda-c180-4aa9-a831-a0840d495fa8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.479501 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-certs\") pod \"machine-config-server-cfzn7\" (UID: \"061a736b-3663-4a73-9836-19724fc4bb71\") " pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.488108 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.507478 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.513391 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-config\") pod \"service-ca-operator-777779d784-6zpzc\" (UID: \"51b2c452-70bd-4a3e-bd74-c7cd6513fd45\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.528622 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.537250 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-serving-cert\") pod \"service-ca-operator-777779d784-6zpzc\" (UID: \"51b2c452-70bd-4a3e-bd74-c7cd6513fd45\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.548595 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.564582 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/061a736b-3663-4a73-9836-19724fc4bb71-node-bootstrap-token\") pod \"machine-config-server-cfzn7\" (UID: \"061a736b-3663-4a73-9836-19724fc4bb71\") " pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.568075 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: W0122 11:42:42.573479 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fb7b4d4_0441_43c4_9596_4d38b369d661.slice/crio-dfe1aa3aa59fbea3da4a9f6c5658128185f20e5c711d45984fad83212361817f WatchSource:0}: Error finding container dfe1aa3aa59fbea3da4a9f6c5658128185f20e5c711d45984fad83212361817f: Status 404 returned error can't find the container with id dfe1aa3aa59fbea3da4a9f6c5658128185f20e5c711d45984fad83212361817f Jan 22 11:42:42 crc kubenswrapper[4874]: W0122 11:42:42.574500 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bec261f_fd1e_44f7_a402_cae34f722b6c.slice/crio-ade3f71032e555ef85241c901460891b890f50c924ff1e07590f13a11e915f4d WatchSource:0}: Error finding container ade3f71032e555ef85241c901460891b890f50c924ff1e07590f13a11e915f4d: Status 404 returned error can't find the container with id ade3f71032e555ef85241c901460891b890f50c924ff1e07590f13a11e915f4d Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.612290 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.631513 4874 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.647763 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.668803 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.687283 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.707824 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.727967 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.748048 4874 request.go:700] Waited for 1.925234271s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Ddefault-dockercfg-2llfx&limit=500&resourceVersion=0 Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.748748 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" event={"ID":"5fb7b4d4-0441-43c4-9596-4d38b369d661","Type":"ContainerStarted","Data":"ba72096e6e9776585fd6ae533977b340482846d49d5ec7499f9db5b281e7bd91"} Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.748788 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" event={"ID":"5fb7b4d4-0441-43c4-9596-4d38b369d661","Type":"ContainerStarted","Data":"dfe1aa3aa59fbea3da4a9f6c5658128185f20e5c711d45984fad83212361817f"} Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.750902 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.751379 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" event={"ID":"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d","Type":"ContainerStarted","Data":"5f0f0f78b2e78150f4a7126b5457975a21b1e7da3a542e7d9eb284ddb9612d7f"} Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.751484 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" event={"ID":"0d49cfd6-9c1b-481f-a9d3-07a99661cf9d","Type":"ContainerStarted","Data":"50cb5bac57621b81371da75029472795c61518f97e75a563aeb7b8f76bccb9ca"} Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.752616 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" event={"ID":"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4","Type":"ContainerStarted","Data":"9c364901ef224c02f1d83fab863948c744f3e6bdc3501b8cd95f34c82feeff08"} Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.752651 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" event={"ID":"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4","Type":"ContainerStarted","Data":"c9acde3f2bdd745a092777b2d7badcee478c91b536adde065c962ce05336fe43"} Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.753507 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" event={"ID":"ca44490a-7fc8-478d-a6e2-670f49816b81","Type":"ContainerStarted","Data":"bddb32e3c72065aeabce6eb546394e141aac0a1229a0392b581610208d4bd605"} Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.754566 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" event={"ID":"9bec261f-fd1e-44f7-a402-cae34f722b6c","Type":"ContainerStarted","Data":"d7e2aada04e2e1a91a486222e031fd03b0a873ea166c39ee1740647dbdb35646"} Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.754597 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" event={"ID":"9bec261f-fd1e-44f7-a402-cae34f722b6c","Type":"ContainerStarted","Data":"ade3f71032e555ef85241c901460891b890f50c924ff1e07590f13a11e915f4d"} Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.755510 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" event={"ID":"f5464a23-ec80-4717-bfe0-6efeab811853","Type":"ContainerStarted","Data":"fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5"} Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.755536 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" event={"ID":"f5464a23-ec80-4717-bfe0-6efeab811853","Type":"ContainerStarted","Data":"d437fa062da853d4db086da0d6818e76b56a3a9b163a7e48bb221915ade96982"} Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.755731 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.757373 4874 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-swcnq container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.757438 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" podUID="f5464a23-ec80-4717-bfe0-6efeab811853" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.767539 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.788315 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.823780 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2xwq\" (UniqueName: \"kubernetes.io/projected/308c337e-1e28-4e34-9ccb-8ae546eee089-kube-api-access-j2xwq\") pod \"cluster-samples-operator-665b6dd947-qfcsk\" (UID: \"308c337e-1e28-4e34-9ccb-8ae546eee089\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.841335 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kg7v\" (UniqueName: \"kubernetes.io/projected/16747491-2b7d-4cb1-841d-61c6f366cf8a-kube-api-access-2kg7v\") pod \"openshift-config-operator-7777fb866f-gv9x5\" (UID: \"16747491-2b7d-4cb1-841d-61c6f366cf8a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.850249 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.871476 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9jtx\" (UniqueName: \"kubernetes.io/projected/c045a558-1990-432e-9965-e918f60aba14-kube-api-access-b9jtx\") pod \"machine-config-controller-84d6567774-dprzc\" (UID: \"c045a558-1990-432e-9965-e918f60aba14\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.879870 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.883937 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh722\" (UniqueName: \"kubernetes.io/projected/304c66b8-6187-47ac-9c57-235c634eaae4-kube-api-access-sh722\") pod \"migrator-59844c95c7-rwcrv\" (UID: \"304c66b8-6187-47ac-9c57-235c634eaae4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.901668 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rprpc\" (UniqueName: \"kubernetes.io/projected/2dad3db6-cddd-457d-8efa-908257ef7cc5-kube-api-access-rprpc\") pod \"console-f9d7485db-wws2s\" (UID: \"2dad3db6-cddd-457d-8efa-908257ef7cc5\") " pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.903281 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.925667 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.926774 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj795\" (UniqueName: \"kubernetes.io/projected/3b9983ba-ed8d-4654-ba74-f25433aa7ee7-kube-api-access-gj795\") pod \"downloads-7954f5f757-jgw4c\" (UID: \"3b9983ba-ed8d-4654-ba74-f25433aa7ee7\") " pod="openshift-console/downloads-7954f5f757-jgw4c" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.944320 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dm8p\" (UniqueName: \"kubernetes.io/projected/88c9b59f-1809-4252-a8fc-cad965848dc0-kube-api-access-8dm8p\") pod \"machine-config-operator-74547568cd-6lgxn\" (UID: \"88c9b59f-1809-4252-a8fc-cad965848dc0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.968937 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmkdz\" (UniqueName: \"kubernetes.io/projected/288e6358-c74b-4597-8968-726a31365f82-kube-api-access-qmkdz\") pod \"apiserver-76f77b778f-cqkbr\" (UID: \"288e6358-c74b-4597-8968-726a31365f82\") " pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:42 crc kubenswrapper[4874]: I0122 11:42:42.989017 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvhz9\" (UniqueName: \"kubernetes.io/projected/e9446a39-9776-4de8-9137-b8952d336419-kube-api-access-lvhz9\") pod \"router-default-5444994796-5cczs\" (UID: \"e9446a39-9776-4de8-9137-b8952d336419\") " pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.010105 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9gf8\" (UniqueName: \"kubernetes.io/projected/9fd4241f-b523-4d66-bcdb-c3bb691765c9-kube-api-access-h9gf8\") pod \"oauth-openshift-558db77b4-8j9tz\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.043790 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m29d2\" (UniqueName: \"kubernetes.io/projected/65f71c2e-ab34-4d33-905f-609555dab78c-kube-api-access-m29d2\") pod \"route-controller-manager-6576b87f9c-2bn87\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.061715 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk"] Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.067521 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f87cb481-5261-4028-99e2-cd57ff6b61e1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gsfnm\" (UID: \"f87cb481-5261-4028-99e2-cd57ff6b61e1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.084796 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9hxd\" (UniqueName: \"kubernetes.io/projected/68dcdc32-485b-435c-81cc-43be463998bb-kube-api-access-c9hxd\") pod \"openshift-apiserver-operator-796bbdcf4f-dnrhx\" (UID: \"68dcdc32-485b-435c-81cc-43be463998bb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.104661 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5"] Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.104923 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-jgw4c" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.110989 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw8xf\" (UniqueName: \"kubernetes.io/projected/8a67b9a8-ad8a-40e3-955c-53aed07a9140-kube-api-access-cw8xf\") pod \"console-operator-58897d9998-mftbv\" (UID: \"8a67b9a8-ad8a-40e3-955c-53aed07a9140\") " pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.122225 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.124061 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdhbq\" (UniqueName: \"kubernetes.io/projected/1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a-kube-api-access-sdhbq\") pod \"packageserver-d55dfcdfc-6pr5c\" (UID: \"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.124323 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.137200 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc"] Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.144498 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqmnt\" (UniqueName: \"kubernetes.io/projected/8d7ccd8e-549f-4b23-bcec-4a1c14e10478-kube-api-access-lqmnt\") pod \"package-server-manager-789f6589d5-vqrw2\" (UID: \"8d7ccd8e-549f-4b23-bcec-4a1c14e10478\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.159606 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.161498 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b42gj\" (UniqueName: \"kubernetes.io/projected/061a736b-3663-4a73-9836-19724fc4bb71-kube-api-access-b42gj\") pod \"machine-config-server-cfzn7\" (UID: \"061a736b-3663-4a73-9836-19724fc4bb71\") " pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.166941 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv"] Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.172439 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.181445 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhgtt\" (UniqueName: \"kubernetes.io/projected/51b2c452-70bd-4a3e-bd74-c7cd6513fd45-kube-api-access-bhgtt\") pod \"service-ca-operator-777779d784-6zpzc\" (UID: \"51b2c452-70bd-4a3e-bd74-c7cd6513fd45\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.188269 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.203948 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.213959 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9bvp\" (UniqueName: \"kubernetes.io/projected/4370babd-21ea-4c7e-81b9-cdd611127094-kube-api-access-d9bvp\") pod \"etcd-operator-b45778765-sjhjk\" (UID: \"4370babd-21ea-4c7e-81b9-cdd611127094\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.215548 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.225823 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmqf8\" (UniqueName: \"kubernetes.io/projected/8b5afeda-c180-4aa9-a831-a0840d495fa8-kube-api-access-hmqf8\") pod \"catalog-operator-68c6474976-dt8w6\" (UID: \"8b5afeda-c180-4aa9-a831-a0840d495fa8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.232559 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.243733 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf6nt\" (UniqueName: \"kubernetes.io/projected/213d34f5-75cd-459c-9e56-2938fe5e3950-kube-api-access-lf6nt\") pod \"marketplace-operator-79b997595-6dvvn\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.273812 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krt27\" (UniqueName: \"kubernetes.io/projected/1795c220-db74-434e-9111-917ff6d95077-kube-api-access-krt27\") pod \"multus-admission-controller-857f4d67dd-f64ps\" (UID: \"1795c220-db74-434e-9111-917ff6d95077\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.290545 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.292593 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjztm\" (UniqueName: \"kubernetes.io/projected/0a15050a-7cbd-40f6-a656-a68293c0878a-kube-api-access-fjztm\") pod \"collect-profiles-29484690-npncx\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.299300 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.305507 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx7xc\" (UniqueName: \"kubernetes.io/projected/f745af09-93ed-4c94-8411-f14b0aaaf1cf-kube-api-access-lx7xc\") pod \"openshift-controller-manager-operator-756b6f6bc6-vsxx4\" (UID: \"f745af09-93ed-4c94-8411-f14b0aaaf1cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.329290 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d80367da-a8d4-4b76-9b33-5526aad85229-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s4fb2\" (UID: \"d80367da-a8d4-4b76-9b33-5526aad85229\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.343868 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.355164 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7t6g\" (UniqueName: \"kubernetes.io/projected/a3d1224e-a43e-4dca-8ad0-239dc50b6d58-kube-api-access-p7t6g\") pod \"kube-storage-version-migrator-operator-b67b599dd-fmhx8\" (UID: \"a3d1224e-a43e-4dca-8ad0-239dc50b6d58\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.371844 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.376368 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.382963 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.391268 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.416693 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.421744 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.434407 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-cfzn7" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.440578 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.450134 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htst2\" (UniqueName: \"kubernetes.io/projected/4f61e177-ea8e-4f61-8ba2-67906f08a00c-kube-api-access-htst2\") pod \"dns-operator-744455d44c-rx6nv\" (UID: \"4f61e177-ea8e-4f61-8ba2-67906f08a00c\") " pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.451847 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtv7h\" (UniqueName: \"kubernetes.io/projected/574225b6-a476-4591-9be9-ddd94e5281ef-kube-api-access-jtv7h\") pod \"olm-operator-6b444d44fb-b6c5m\" (UID: \"574225b6-a476-4591-9be9-ddd94e5281ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.463802 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489738 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gp99\" (UniqueName: \"kubernetes.io/projected/67454032-f4d2-418e-afad-a48c0a80007d-kube-api-access-6gp99\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489776 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67454032-f4d2-418e-afad-a48c0a80007d-trusted-ca\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489792 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-trusted-ca\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489809 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/90a037cd-cacf-4706-a857-f65c8f16c384-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489829 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a785e748-a557-4c24-8a4d-dc03cfc3c357-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6jqjk\" (UID: \"a785e748-a557-4c24-8a4d-dc03cfc3c357\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489847 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/90a037cd-cacf-4706-a857-f65c8f16c384-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489863 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ee5bb84-2932-4f1c-8af7-39e10c1260c2-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vpjnl\" (UID: \"4ee5bb84-2932-4f1c-8af7-39e10c1260c2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489877 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6a39a4e6-a146-4ff3-8e90-eb516fe27e38-signing-cabundle\") pod \"service-ca-9c57cc56f-dwcjx\" (UID: \"6a39a4e6-a146-4ff3-8e90-eb516fe27e38\") " pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489900 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/67454032-f4d2-418e-afad-a48c0a80007d-metrics-tls\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489915 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6a39a4e6-a146-4ff3-8e90-eb516fe27e38-signing-key\") pod \"service-ca-9c57cc56f-dwcjx\" (UID: \"6a39a4e6-a146-4ff3-8e90-eb516fe27e38\") " pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489945 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489971 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee5bb84-2932-4f1c-8af7-39e10c1260c2-config\") pod \"kube-controller-manager-operator-78b949d7b-vpjnl\" (UID: \"4ee5bb84-2932-4f1c-8af7-39e10c1260c2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.489989 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp7dh\" (UniqueName: \"kubernetes.io/projected/a785e748-a557-4c24-8a4d-dc03cfc3c357-kube-api-access-lp7dh\") pod \"control-plane-machine-set-operator-78cbb6b69f-6jqjk\" (UID: \"a785e748-a557-4c24-8a4d-dc03cfc3c357\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.490006 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-registry-tls\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.490020 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-bound-sa-token\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.490035 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwtlg\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-kube-api-access-zwtlg\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.490085 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ee5bb84-2932-4f1c-8af7-39e10c1260c2-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vpjnl\" (UID: \"4ee5bb84-2932-4f1c-8af7-39e10c1260c2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.490113 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-registry-certificates\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.490155 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rb79\" (UniqueName: \"kubernetes.io/projected/6a39a4e6-a146-4ff3-8e90-eb516fe27e38-kube-api-access-7rb79\") pod \"service-ca-9c57cc56f-dwcjx\" (UID: \"6a39a4e6-a146-4ff3-8e90-eb516fe27e38\") " pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.490188 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67454032-f4d2-418e-afad-a48c0a80007d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: E0122 11:42:43.492373 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:43.992361738 +0000 UTC m=+137.837432808 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.596726 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597167 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67454032-f4d2-418e-afad-a48c0a80007d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597261 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjtx5\" (UniqueName: \"kubernetes.io/projected/41a97907-0515-44cd-b6a4-87463a9b3819-kube-api-access-vjtx5\") pod \"dns-default-kxvzk\" (UID: \"41a97907-0515-44cd-b6a4-87463a9b3819\") " pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597365 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-registration-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597482 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-csi-data-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597554 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl6zw\" (UniqueName: \"kubernetes.io/projected/a54b65f9-405b-4a90-a9eb-e0438722ffa8-kube-api-access-sl6zw\") pod \"ingress-canary-s5xxd\" (UID: \"a54b65f9-405b-4a90-a9eb-e0438722ffa8\") " pod="openshift-ingress-canary/ingress-canary-s5xxd" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597582 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gp99\" (UniqueName: \"kubernetes.io/projected/67454032-f4d2-418e-afad-a48c0a80007d-kube-api-access-6gp99\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597607 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67454032-f4d2-418e-afad-a48c0a80007d-trusted-ca\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597651 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/90a037cd-cacf-4706-a857-f65c8f16c384-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597666 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-trusted-ca\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597680 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a54b65f9-405b-4a90-a9eb-e0438722ffa8-cert\") pod \"ingress-canary-s5xxd\" (UID: \"a54b65f9-405b-4a90-a9eb-e0438722ffa8\") " pod="openshift-ingress-canary/ingress-canary-s5xxd" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597707 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-socket-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597721 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trsjs\" (UniqueName: \"kubernetes.io/projected/7a9c3919-d376-4655-818e-3a88a8c0b883-kube-api-access-trsjs\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597757 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a785e748-a557-4c24-8a4d-dc03cfc3c357-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6jqjk\" (UID: \"a785e748-a557-4c24-8a4d-dc03cfc3c357\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597776 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/90a037cd-cacf-4706-a857-f65c8f16c384-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597800 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6a39a4e6-a146-4ff3-8e90-eb516fe27e38-signing-cabundle\") pod \"service-ca-9c57cc56f-dwcjx\" (UID: \"6a39a4e6-a146-4ff3-8e90-eb516fe27e38\") " pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597815 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ee5bb84-2932-4f1c-8af7-39e10c1260c2-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vpjnl\" (UID: \"4ee5bb84-2932-4f1c-8af7-39e10c1260c2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597864 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/67454032-f4d2-418e-afad-a48c0a80007d-metrics-tls\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597888 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6a39a4e6-a146-4ff3-8e90-eb516fe27e38-signing-key\") pod \"service-ca-9c57cc56f-dwcjx\" (UID: \"6a39a4e6-a146-4ff3-8e90-eb516fe27e38\") " pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597934 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/41a97907-0515-44cd-b6a4-87463a9b3819-metrics-tls\") pod \"dns-default-kxvzk\" (UID: \"41a97907-0515-44cd-b6a4-87463a9b3819\") " pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.597963 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee5bb84-2932-4f1c-8af7-39e10c1260c2-config\") pod \"kube-controller-manager-operator-78b949d7b-vpjnl\" (UID: \"4ee5bb84-2932-4f1c-8af7-39e10c1260c2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.598010 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp7dh\" (UniqueName: \"kubernetes.io/projected/a785e748-a557-4c24-8a4d-dc03cfc3c357-kube-api-access-lp7dh\") pod \"control-plane-machine-set-operator-78cbb6b69f-6jqjk\" (UID: \"a785e748-a557-4c24-8a4d-dc03cfc3c357\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.598037 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-registry-tls\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.598097 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwtlg\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-kube-api-access-zwtlg\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.598119 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-plugins-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.598168 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-bound-sa-token\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.598227 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ee5bb84-2932-4f1c-8af7-39e10c1260c2-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vpjnl\" (UID: \"4ee5bb84-2932-4f1c-8af7-39e10c1260c2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.598284 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-registry-certificates\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.598317 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41a97907-0515-44cd-b6a4-87463a9b3819-config-volume\") pod \"dns-default-kxvzk\" (UID: \"41a97907-0515-44cd-b6a4-87463a9b3819\") " pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.598364 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rb79\" (UniqueName: \"kubernetes.io/projected/6a39a4e6-a146-4ff3-8e90-eb516fe27e38-kube-api-access-7rb79\") pod \"service-ca-9c57cc56f-dwcjx\" (UID: \"6a39a4e6-a146-4ff3-8e90-eb516fe27e38\") " pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.598418 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-mountpoint-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: E0122 11:42:43.598653 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:44.098616278 +0000 UTC m=+137.943687358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.599225 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/90a037cd-cacf-4706-a857-f65c8f16c384-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.599938 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-trusted-ca\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.604370 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-registry-certificates\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.607314 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67454032-f4d2-418e-afad-a48c0a80007d-trusted-ca\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.608161 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee5bb84-2932-4f1c-8af7-39e10c1260c2-config\") pod \"kube-controller-manager-operator-78b949d7b-vpjnl\" (UID: \"4ee5bb84-2932-4f1c-8af7-39e10c1260c2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.610704 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a785e748-a557-4c24-8a4d-dc03cfc3c357-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6jqjk\" (UID: \"a785e748-a557-4c24-8a4d-dc03cfc3c357\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.613694 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6a39a4e6-a146-4ff3-8e90-eb516fe27e38-signing-cabundle\") pod \"service-ca-9c57cc56f-dwcjx\" (UID: \"6a39a4e6-a146-4ff3-8e90-eb516fe27e38\") " pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.613931 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.622967 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6a39a4e6-a146-4ff3-8e90-eb516fe27e38-signing-key\") pod \"service-ca-9c57cc56f-dwcjx\" (UID: \"6a39a4e6-a146-4ff3-8e90-eb516fe27e38\") " pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.629661 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.640370 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67454032-f4d2-418e-afad-a48c0a80007d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.640898 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rb79\" (UniqueName: \"kubernetes.io/projected/6a39a4e6-a146-4ff3-8e90-eb516fe27e38-kube-api-access-7rb79\") pod \"service-ca-9c57cc56f-dwcjx\" (UID: \"6a39a4e6-a146-4ff3-8e90-eb516fe27e38\") " pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.644691 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/90a037cd-cacf-4706-a857-f65c8f16c384-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.648123 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/67454032-f4d2-418e-afad-a48c0a80007d-metrics-tls\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.659263 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-registry-tls\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.665119 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwtlg\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-kube-api-access-zwtlg\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.667759 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp7dh\" (UniqueName: \"kubernetes.io/projected/a785e748-a557-4c24-8a4d-dc03cfc3c357-kube-api-access-lp7dh\") pod \"control-plane-machine-set-operator-78cbb6b69f-6jqjk\" (UID: \"a785e748-a557-4c24-8a4d-dc03cfc3c357\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.671971 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ee5bb84-2932-4f1c-8af7-39e10c1260c2-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vpjnl\" (UID: \"4ee5bb84-2932-4f1c-8af7-39e10c1260c2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.673749 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-jgw4c"] Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.686786 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gp99\" (UniqueName: \"kubernetes.io/projected/67454032-f4d2-418e-afad-a48c0a80007d-kube-api-access-6gp99\") pod \"ingress-operator-5b745b69d9-89ldw\" (UID: \"67454032-f4d2-418e-afad-a48c0a80007d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.696920 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ee5bb84-2932-4f1c-8af7-39e10c1260c2-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vpjnl\" (UID: \"4ee5bb84-2932-4f1c-8af7-39e10c1260c2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.702003 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-bound-sa-token\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.702312 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703320 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703365 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/41a97907-0515-44cd-b6a4-87463a9b3819-metrics-tls\") pod \"dns-default-kxvzk\" (UID: \"41a97907-0515-44cd-b6a4-87463a9b3819\") " pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703390 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-plugins-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703437 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41a97907-0515-44cd-b6a4-87463a9b3819-config-volume\") pod \"dns-default-kxvzk\" (UID: \"41a97907-0515-44cd-b6a4-87463a9b3819\") " pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703464 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-mountpoint-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703494 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjtx5\" (UniqueName: \"kubernetes.io/projected/41a97907-0515-44cd-b6a4-87463a9b3819-kube-api-access-vjtx5\") pod \"dns-default-kxvzk\" (UID: \"41a97907-0515-44cd-b6a4-87463a9b3819\") " pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703515 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-registration-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703544 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-csi-data-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703571 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl6zw\" (UniqueName: \"kubernetes.io/projected/a54b65f9-405b-4a90-a9eb-e0438722ffa8-kube-api-access-sl6zw\") pod \"ingress-canary-s5xxd\" (UID: \"a54b65f9-405b-4a90-a9eb-e0438722ffa8\") " pod="openshift-ingress-canary/ingress-canary-s5xxd" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703592 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a54b65f9-405b-4a90-a9eb-e0438722ffa8-cert\") pod \"ingress-canary-s5xxd\" (UID: \"a54b65f9-405b-4a90-a9eb-e0438722ffa8\") " pod="openshift-ingress-canary/ingress-canary-s5xxd" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703614 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-socket-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.703633 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trsjs\" (UniqueName: \"kubernetes.io/projected/7a9c3919-d376-4655-818e-3a88a8c0b883-kube-api-access-trsjs\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: E0122 11:42:43.704045 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:44.204030691 +0000 UTC m=+138.049101761 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.704839 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-plugins-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.704930 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-csi-data-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.705221 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-mountpoint-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.705365 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-registration-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.705440 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7a9c3919-d376-4655-818e-3a88a8c0b883-socket-dir\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.706368 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41a97907-0515-44cd-b6a4-87463a9b3819-config-volume\") pod \"dns-default-kxvzk\" (UID: \"41a97907-0515-44cd-b6a4-87463a9b3819\") " pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.709099 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mftbv"] Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.713116 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a54b65f9-405b-4a90-a9eb-e0438722ffa8-cert\") pod \"ingress-canary-s5xxd\" (UID: \"a54b65f9-405b-4a90-a9eb-e0438722ffa8\") " pod="openshift-ingress-canary/ingress-canary-s5xxd" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.730757 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/41a97907-0515-44cd-b6a4-87463a9b3819-metrics-tls\") pod \"dns-default-kxvzk\" (UID: \"41a97907-0515-44cd-b6a4-87463a9b3819\") " pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.749645 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trsjs\" (UniqueName: \"kubernetes.io/projected/7a9c3919-d376-4655-818e-3a88a8c0b883-kube-api-access-trsjs\") pod \"csi-hostpathplugin-l87tw\" (UID: \"7a9c3919-d376-4655-818e-3a88a8c0b883\") " pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.750077 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.761064 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjtx5\" (UniqueName: \"kubernetes.io/projected/41a97907-0515-44cd-b6a4-87463a9b3819-kube-api-access-vjtx5\") pod \"dns-default-kxvzk\" (UID: \"41a97907-0515-44cd-b6a4-87463a9b3819\") " pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.792774 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl6zw\" (UniqueName: \"kubernetes.io/projected/a54b65f9-405b-4a90-a9eb-e0438722ffa8-kube-api-access-sl6zw\") pod \"ingress-canary-s5xxd\" (UID: \"a54b65f9-405b-4a90-a9eb-e0438722ffa8\") " pod="openshift-ingress-canary/ingress-canary-s5xxd" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.795227 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8j9tz"] Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.797926 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-l87tw" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.804561 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:43 crc kubenswrapper[4874]: E0122 11:42:43.804976 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:44.304960374 +0000 UTC m=+138.150031444 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.808607 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.816358 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" event={"ID":"c045a558-1990-432e-9965-e918f60aba14","Type":"ContainerStarted","Data":"8467f2a46fbebb9e4dc1def39d91236ab29caa6aee23ce46778cfaaf0d83ef57"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.816417 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" event={"ID":"c045a558-1990-432e-9965-e918f60aba14","Type":"ContainerStarted","Data":"5017eeb33b0ae024aedc6a079168c60ce6127106057a25534829aeb4246ab6ea"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.830700 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" event={"ID":"308c337e-1e28-4e34-9ccb-8ae546eee089","Type":"ContainerStarted","Data":"752c9a819a1d05cac2b0fb64e0acd5937788e935466322f6a1fed0a7fe7f456a"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.830762 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" event={"ID":"308c337e-1e28-4e34-9ccb-8ae546eee089","Type":"ContainerStarted","Data":"c0027027d3937714e90c671faaa05115d50098db457aa8f62164545362bdefe2"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.831760 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-cfzn7" event={"ID":"061a736b-3663-4a73-9836-19724fc4bb71","Type":"ContainerStarted","Data":"a1e0b7129a01f83cf10c32190ce32afcee33053723d51469859d7cc38a7fe962"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.833231 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5cczs" event={"ID":"e9446a39-9776-4de8-9137-b8952d336419","Type":"ContainerStarted","Data":"9ebf022f68edda676021cc0fbeb0a68174f6ad4fbc2e4ef2d9bb751280d4ec65"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.833277 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5cczs" event={"ID":"e9446a39-9776-4de8-9137-b8952d336419","Type":"ContainerStarted","Data":"1fdc6176317759ddd858337b2ed6123236c395a7fb1b03f05858040cdad255cb"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.846519 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv" event={"ID":"304c66b8-6187-47ac-9c57-235c634eaae4","Type":"ContainerStarted","Data":"a4a37b7be2b2f997a241b5ad06f4719213fa43e63628783dad8f0e263be8b6f5"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.846565 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv" event={"ID":"304c66b8-6187-47ac-9c57-235c634eaae4","Type":"ContainerStarted","Data":"220af10eecb4f8ff7f55663bdec7f3f351e2a96907240f1f43525f975f7ee9e2"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.858232 4874 generic.go:334] "Generic (PLEG): container finished" podID="16747491-2b7d-4cb1-841d-61c6f366cf8a" containerID="72a7a981031cece1c7f7c83a7047c6627dc5039ed7b07331ef32054a2f4fe2ca" exitCode=0 Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.858290 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" event={"ID":"16747491-2b7d-4cb1-841d-61c6f366cf8a","Type":"ContainerDied","Data":"72a7a981031cece1c7f7c83a7047c6627dc5039ed7b07331ef32054a2f4fe2ca"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.858311 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" event={"ID":"16747491-2b7d-4cb1-841d-61c6f366cf8a","Type":"ContainerStarted","Data":"b8aa1c4a3f08c4320af86054c2394b97a51506fdbd487a72ccf2be7d5442dde2"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.870021 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" event={"ID":"6c9561fa-20b6-4f87-aacc-cf0e0665ffa4","Type":"ContainerStarted","Data":"db5d00e60244a0cbee5256d6ffedf435df83bed45bda5a66f991aeebaf5871e4"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.871912 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c"] Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.873848 4874 generic.go:334] "Generic (PLEG): container finished" podID="ca44490a-7fc8-478d-a6e2-670f49816b81" containerID="0d06809c08191501d88b6b9c6ff0fe96ceb607cba26a2b0224d79047b5a8139f" exitCode=0 Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.873898 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" event={"ID":"ca44490a-7fc8-478d-a6e2-670f49816b81","Type":"ContainerDied","Data":"0d06809c08191501d88b6b9c6ff0fe96ceb607cba26a2b0224d79047b5a8139f"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.877770 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" event={"ID":"9bec261f-fd1e-44f7-a402-cae34f722b6c","Type":"ContainerStarted","Data":"fcf3b93bfefbdbe8ae3094c95bd63cf0c0416ba9835712bcfbc9f24cde2e379b"} Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.881602 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s5xxd" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.893899 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.907897 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:43 crc kubenswrapper[4874]: E0122 11:42:43.908230 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:44.4082196 +0000 UTC m=+138.253290670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.920467 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.923227 4874 csr.go:261] certificate signing request csr-x4fcz is approved, waiting to be issued Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.929048 4874 csr.go:257] certificate signing request csr-x4fcz is issued Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.952605 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk" Jan 22 11:42:43 crc kubenswrapper[4874]: I0122 11:42:43.962314 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.014167 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.015426 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:44.515411459 +0000 UTC m=+138.360482529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.116314 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.116850 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:44.616827437 +0000 UTC m=+138.461898507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.202943 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" podStartSLOduration=120.202921591 podStartE2EDuration="2m0.202921591s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:44.201667242 +0000 UTC m=+138.046738402" watchObservedRunningTime="2026-01-22 11:42:44.202921591 +0000 UTC m=+138.047992651" Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.218053 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.218465 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:44.718424931 +0000 UTC m=+138.563496011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.218491 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.218756 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:44.718749571 +0000 UTC m=+138.563820641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.233953 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.319617 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.320139 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:44.820124567 +0000 UTC m=+138.665195637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.349599 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ssbbs" podStartSLOduration=120.349572226 podStartE2EDuration="2m0.349572226s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:44.347994235 +0000 UTC m=+138.193065305" watchObservedRunningTime="2026-01-22 11:42:44.349572226 +0000 UTC m=+138.194643296" Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.424115 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.424656 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:44.924642312 +0000 UTC m=+138.769713382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.443966 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:44 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:44 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:44 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.444017 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.526165 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.526325 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.026296768 +0000 UTC m=+138.871367838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.526389 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.526700 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.02669143 +0000 UTC m=+138.871762560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.587981 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmwcz" podStartSLOduration=119.587960212 podStartE2EDuration="1m59.587960212s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:44.539189564 +0000 UTC m=+138.384260634" watchObservedRunningTime="2026-01-22 11:42:44.587960212 +0000 UTC m=+138.433031272" Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.627336 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.627537 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.127509479 +0000 UTC m=+138.972580559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.629011 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.629342 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.129330716 +0000 UTC m=+138.974401786 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.661491 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-869fn" podStartSLOduration=120.661469989 podStartE2EDuration="2m0.661469989s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:44.661317835 +0000 UTC m=+138.506388925" watchObservedRunningTime="2026-01-22 11:42:44.661469989 +0000 UTC m=+138.506541069" Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.706419 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-5cczs" podStartSLOduration=120.706368355 podStartE2EDuration="2m0.706368355s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:44.703970609 +0000 UTC m=+138.549041679" watchObservedRunningTime="2026-01-22 11:42:44.706368355 +0000 UTC m=+138.551439425" Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.730483 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.730723 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.230691942 +0000 UTC m=+139.075763012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.730817 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.731338 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.231330583 +0000 UTC m=+139.076401653 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.832628 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.832833 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.332800221 +0000 UTC m=+139.177871291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.832920 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.833378 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.333370519 +0000 UTC m=+139.178441589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.883442 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" event={"ID":"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a","Type":"ContainerStarted","Data":"3826d56a62c5fd00e3561a10b809db8a4a5499338a4f0cc48cdc80320d70af51"} Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.884329 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" event={"ID":"9fd4241f-b523-4d66-bcdb-c3bb691765c9","Type":"ContainerStarted","Data":"ebdd6071b5dc3f0f50c15dd7a89f7157d85c2d83e8eb328a3a8f9e41e0ceff64"} Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.885541 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mftbv" event={"ID":"8a67b9a8-ad8a-40e3-955c-53aed07a9140","Type":"ContainerStarted","Data":"f0c2f029d63ce8d982a1f81d10a536666090b310e7796537ca9f84d3c4a28a3a"} Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.889241 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-jgw4c" event={"ID":"3b9983ba-ed8d-4654-ba74-f25433aa7ee7","Type":"ContainerStarted","Data":"5b427dcc0615916e55edeb38400ee108b23bccc5cf915114cd4663a0326595ae"} Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.930695 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-22 11:37:43 +0000 UTC, rotation deadline is 2026-11-27 11:04:23.203483956 +0000 UTC Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.930740 4874 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7415h21m38.272747277s for next certificate rotation Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.934335 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.934632 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.434619872 +0000 UTC m=+139.279690942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:44 crc kubenswrapper[4874]: I0122 11:42:44.934731 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:44 crc kubenswrapper[4874]: E0122 11:42:44.935045 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.435037475 +0000 UTC m=+139.280108545 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.036845 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:45 crc kubenswrapper[4874]: E0122 11:42:45.037992 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.537977001 +0000 UTC m=+139.383048071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.138341 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:45 crc kubenswrapper[4874]: E0122 11:42:45.138903 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.638883742 +0000 UTC m=+139.483954812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.241051 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:45 crc kubenswrapper[4874]: E0122 11:42:45.241326 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.741302841 +0000 UTC m=+139.586373911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.244432 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:45 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:45 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:45 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.244483 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.254851 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7h7c7" podStartSLOduration=121.254829038 podStartE2EDuration="2m1.254829038s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:45.206469353 +0000 UTC m=+139.051540413" watchObservedRunningTime="2026-01-22 11:42:45.254829038 +0000 UTC m=+139.099900118" Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.342631 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:45 crc kubenswrapper[4874]: E0122 11:42:45.343225 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.843197624 +0000 UTC m=+139.688268694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.357504 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wws2s"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.384930 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.418854 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.451161 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:45 crc kubenswrapper[4874]: E0122 11:42:45.451638 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:45.951615112 +0000 UTC m=+139.796686182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.463111 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2"] Jan 22 11:42:45 crc kubenswrapper[4874]: W0122 11:42:45.518990 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd80367da_a8d4_4b76_9b33_5526aad85229.slice/crio-aa7fd2628587b5bf777ed91fb7b1e43479d8ee301c2b3642663b5af0e306946b WatchSource:0}: Error finding container aa7fd2628587b5bf777ed91fb7b1e43479d8ee301c2b3642663b5af0e306946b: Status 404 returned error can't find the container with id aa7fd2628587b5bf777ed91fb7b1e43479d8ee301c2b3642663b5af0e306946b Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.552483 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:45 crc kubenswrapper[4874]: E0122 11:42:45.552893 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:46.052881165 +0000 UTC m=+139.897952225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.654125 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:45 crc kubenswrapper[4874]: E0122 11:42:45.654749 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:46.154725917 +0000 UTC m=+139.999796987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.739314 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-sjhjk"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.745049 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6dvvn"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.755761 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:45 crc kubenswrapper[4874]: E0122 11:42:45.756719 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:46.256693851 +0000 UTC m=+140.101764921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.769363 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.788379 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc"] Jan 22 11:42:45 crc kubenswrapper[4874]: W0122 11:42:45.795619 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4370babd_21ea_4c7e_81b9_cdd611127094.slice/crio-76be14ef4c33b248a1626a35b1c81885884c2b29b18db891810d8b4924480d1f WatchSource:0}: Error finding container 76be14ef4c33b248a1626a35b1c81885884c2b29b18db891810d8b4924480d1f: Status 404 returned error can't find the container with id 76be14ef4c33b248a1626a35b1c81885884c2b29b18db891810d8b4924480d1f Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.820086 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rx6nv"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.839657 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s5xxd"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.855984 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.858910 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:45 crc kubenswrapper[4874]: E0122 11:42:45.859362 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:46.359346349 +0000 UTC m=+140.204417419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.872134 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.888116 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.888998 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-f64ps"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.921602 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv" event={"ID":"304c66b8-6187-47ac-9c57-235c634eaae4","Type":"ContainerStarted","Data":"e56f4c7d4db0bae15a14fc7b2128ccfcbc9d52b63b70f91553c634c3854c88f2"} Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.945440 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.948415 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" event={"ID":"c045a558-1990-432e-9965-e918f60aba14","Type":"ContainerStarted","Data":"08ce10fae4bc524ea5d487aa15edc7a8c96be9cf3039bbe3b8438502f610d1c1"} Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.951047 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wws2s" event={"ID":"2dad3db6-cddd-457d-8efa-908257ef7cc5","Type":"ContainerStarted","Data":"eb821e05f105388a16d1f483490ba591cece2cfd3ac61a1d7e8d081cfcb2eed3"} Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.951082 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wws2s" event={"ID":"2dad3db6-cddd-457d-8efa-908257ef7cc5","Type":"ContainerStarted","Data":"65a9eee8d0422a080a2e9b23d35a87bc7b70cf7c5c56b2bd7476e73873a4352e"} Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.961517 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6"] Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.962082 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:45 crc kubenswrapper[4874]: E0122 11:42:45.962556 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:46.462534592 +0000 UTC m=+140.307605662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:45 crc kubenswrapper[4874]: I0122 11:42:45.989265 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm"] Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:45.997023 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m"] Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.001652 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx"] Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.008948 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-jgw4c" event={"ID":"3b9983ba-ed8d-4654-ba74-f25433aa7ee7","Type":"ContainerStarted","Data":"9657bc5011b0e6d3cba42388e839ce8c1ef84788483d8f44ea92cb538f0db28d"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.009175 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-l87tw"] Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.009206 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-jgw4c" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.021683 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" event={"ID":"213d34f5-75cd-459c-9e56-2938fe5e3950","Type":"ContainerStarted","Data":"e55fae314e5ccbaff31d4019382891789d84d2a8792b6ac6cab4bdcc20e97e1e"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.021717 4874 patch_prober.go:28] interesting pod/downloads-7954f5f757-jgw4c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.021830 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jgw4c" podUID="3b9983ba-ed8d-4654-ba74-f25433aa7ee7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.033667 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-cqkbr"] Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.039423 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4"] Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.039466 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" event={"ID":"88c9b59f-1809-4252-a8fc-cad965848dc0","Type":"ContainerStarted","Data":"3ecbbc38ae38ecf2ffe8c819f8e71acbf0b4c3670a1bdce34ce15e80af22e2cf"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.039491 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" event={"ID":"88c9b59f-1809-4252-a8fc-cad965848dc0","Type":"ContainerStarted","Data":"bd0be220935c39fe411c3f192ea4860e097c146a3b74e883c05c2b278e2ae44e"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.058658 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s5xxd" event={"ID":"a54b65f9-405b-4a90-a9eb-e0438722ffa8","Type":"ContainerStarted","Data":"908285d52bb7b2261e2614aca0390cbf755acb3b0e494795250ea03447cbb38a"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.063148 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:46 crc kubenswrapper[4874]: E0122 11:42:46.065018 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:46.564995102 +0000 UTC m=+140.410066212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.071169 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2"] Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.071283 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kxvzk"] Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.073560 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dwcjx"] Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.074379 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rwcrv" podStartSLOduration=122.074357078 podStartE2EDuration="2m2.074357078s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:45.986331102 +0000 UTC m=+139.831402182" watchObservedRunningTime="2026-01-22 11:42:46.074357078 +0000 UTC m=+139.919428148" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.074736 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" event={"ID":"16747491-2b7d-4cb1-841d-61c6f366cf8a","Type":"ContainerStarted","Data":"71a3b365a76cdda3ccbd96b1bfcbd8fc5a75457ba1b881350b13e9a0f2656599"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.074814 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.090483 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-wws2s" podStartSLOduration=122.090460975 podStartE2EDuration="2m2.090460975s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.033509599 +0000 UTC m=+139.878580669" watchObservedRunningTime="2026-01-22 11:42:46.090460975 +0000 UTC m=+139.935532065" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.115648 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" event={"ID":"308c337e-1e28-4e34-9ccb-8ae546eee089","Type":"ContainerStarted","Data":"be463a8baa10fb0031e37d1e347f370601b34a7e60a08b7f71e1880c31d503a3"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.128521 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dprzc" podStartSLOduration=121.128494094 podStartE2EDuration="2m1.128494094s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.064687022 +0000 UTC m=+139.909758102" watchObservedRunningTime="2026-01-22 11:42:46.128494094 +0000 UTC m=+139.973565164" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.129712 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-jgw4c" podStartSLOduration=122.129705853 podStartE2EDuration="2m2.129705853s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.111097976 +0000 UTC m=+139.956169036" watchObservedRunningTime="2026-01-22 11:42:46.129705853 +0000 UTC m=+139.974776933" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.149601 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" event={"ID":"4370babd-21ea-4c7e-81b9-cdd611127094","Type":"ContainerStarted","Data":"76be14ef4c33b248a1626a35b1c81885884c2b29b18db891810d8b4924480d1f"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.151684 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" podStartSLOduration=122.151667525 podStartE2EDuration="2m2.151667525s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.1499197 +0000 UTC m=+139.994990780" watchObservedRunningTime="2026-01-22 11:42:46.151667525 +0000 UTC m=+139.996738595" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.168828 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" event={"ID":"9fd4241f-b523-4d66-bcdb-c3bb691765c9","Type":"ContainerStarted","Data":"0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.169827 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.169949 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:46 crc kubenswrapper[4874]: E0122 11:42:46.170531 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:46.670518659 +0000 UTC m=+140.515589809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.183537 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" event={"ID":"4ee5bb84-2932-4f1c-8af7-39e10c1260c2","Type":"ContainerStarted","Data":"487cdb5fd2da997d6662b515aa393cd39aee1d829581faded7a8209d4fd7ea1f"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.184368 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qfcsk" podStartSLOduration=122.184350965 podStartE2EDuration="2m2.184350965s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.183382475 +0000 UTC m=+140.028453545" watchObservedRunningTime="2026-01-22 11:42:46.184350965 +0000 UTC m=+140.029422035" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.195363 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" event={"ID":"51b2c452-70bd-4a3e-bd74-c7cd6513fd45","Type":"ContainerStarted","Data":"ac795fc6eacd1d28ef567fbec006e021c43ea9f8710edbc2ea7f75db77c4337c"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.197587 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" event={"ID":"1cc3bb5c-983c-4a20-bbb1-2bb0bd24cd5a","Type":"ContainerStarted","Data":"52cecb435816c6b6693584d7305d4bc7b0c6e9547728c405a0915d37d36f1e64"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.198707 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.210959 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-cfzn7" event={"ID":"061a736b-3663-4a73-9836-19724fc4bb71","Type":"ContainerStarted","Data":"51ab62069e8e04efad5817dfc014b1805c7af17e23c9c681963c871384a4a77c"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.235291 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" podStartSLOduration=122.23525133 podStartE2EDuration="2m2.23525133s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.233823156 +0000 UTC m=+140.078894236" watchObservedRunningTime="2026-01-22 11:42:46.23525133 +0000 UTC m=+140.080322400" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.241931 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" event={"ID":"ca44490a-7fc8-478d-a6e2-670f49816b81","Type":"ContainerStarted","Data":"636e474851c2358938ca9c20c5d6b3952ed505125743dd5a92b549bbb459369a"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.244340 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:46 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:46 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:46 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.244387 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.251771 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mftbv" event={"ID":"8a67b9a8-ad8a-40e3-955c-53aed07a9140","Type":"ContainerStarted","Data":"b16db8afdc1fe803ec4b2fa8f1b30642d382469740f350302bd35307c74b82ad"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.252417 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.263431 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" event={"ID":"d80367da-a8d4-4b76-9b33-5526aad85229","Type":"ContainerStarted","Data":"aa7fd2628587b5bf777ed91fb7b1e43479d8ee301c2b3642663b5af0e306946b"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.263718 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" podStartSLOduration=121.263694997 podStartE2EDuration="2m1.263694997s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.262884281 +0000 UTC m=+140.107955351" watchObservedRunningTime="2026-01-22 11:42:46.263694997 +0000 UTC m=+140.108766067" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.271406 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:46 crc kubenswrapper[4874]: E0122 11:42:46.272732 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:46.772716752 +0000 UTC m=+140.617787822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.276598 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" event={"ID":"68dcdc32-485b-435c-81cc-43be463998bb","Type":"ContainerStarted","Data":"e85515ccbecf22ae2317d95af2bacdfd7374fc2ea6b1a492248ab99963ad1934"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.276650 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" event={"ID":"68dcdc32-485b-435c-81cc-43be463998bb","Type":"ContainerStarted","Data":"510bfb39c5dbd64db543d1bf41a991a9cd9ebe08f2eb0c92a25a2e8f9ee5a810"} Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.316904 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" podStartSLOduration=122.316868704 podStartE2EDuration="2m2.316868704s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.315832102 +0000 UTC m=+140.160903182" watchObservedRunningTime="2026-01-22 11:42:46.316868704 +0000 UTC m=+140.161939774" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.317899 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-cfzn7" podStartSLOduration=6.317882706 podStartE2EDuration="6.317882706s" podCreationTimestamp="2026-01-22 11:42:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.296991957 +0000 UTC m=+140.142063027" watchObservedRunningTime="2026-01-22 11:42:46.317882706 +0000 UTC m=+140.162953776" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.348318 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" podStartSLOduration=121.348301545 podStartE2EDuration="2m1.348301545s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.344745093 +0000 UTC m=+140.189816183" watchObservedRunningTime="2026-01-22 11:42:46.348301545 +0000 UTC m=+140.193372615" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.376239 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dnrhx" podStartSLOduration=122.376211295 podStartE2EDuration="2m2.376211295s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.375813862 +0000 UTC m=+140.220884932" watchObservedRunningTime="2026-01-22 11:42:46.376211295 +0000 UTC m=+140.221282365" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.376307 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:46 crc kubenswrapper[4874]: E0122 11:42:46.384414 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:46.884378062 +0000 UTC m=+140.729449132 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.423983 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-mftbv" podStartSLOduration=122.423964011 podStartE2EDuration="2m2.423964011s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:46.421588346 +0000 UTC m=+140.266659416" watchObservedRunningTime="2026-01-22 11:42:46.423964011 +0000 UTC m=+140.269035081" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.477548 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:46 crc kubenswrapper[4874]: E0122 11:42:46.477982 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:46.977961914 +0000 UTC m=+140.823032994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.583139 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:46 crc kubenswrapper[4874]: E0122 11:42:46.583541 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:47.083525182 +0000 UTC m=+140.928596252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.686216 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:46 crc kubenswrapper[4874]: E0122 11:42:46.686687 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:47.186653863 +0000 UTC m=+141.031724923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.789579 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:46 crc kubenswrapper[4874]: E0122 11:42:46.790228 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:47.290216678 +0000 UTC m=+141.135287758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.866493 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.866899 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.880956 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6pr5c" Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.890847 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:46 crc kubenswrapper[4874]: E0122 11:42:46.891279 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:47.391258694 +0000 UTC m=+141.236329774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:46 crc kubenswrapper[4874]: I0122 11:42:46.992481 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:46 crc kubenswrapper[4874]: E0122 11:42:46.992908 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:47.492891219 +0000 UTC m=+141.337962289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.102439 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:47 crc kubenswrapper[4874]: E0122 11:42:47.102988 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:47.602967229 +0000 UTC m=+141.448038299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.170514 4874 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8j9tz container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.170588 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" podUID="9fd4241f-b523-4d66-bcdb-c3bb691765c9" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.207034 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:47 crc kubenswrapper[4874]: E0122 11:42:47.228493 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:47.728475997 +0000 UTC m=+141.573547067 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.236738 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:47 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:47 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:47 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.236831 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.253678 4874 patch_prober.go:28] interesting pod/console-operator-58897d9998-mftbv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.253930 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-mftbv" podUID="8a67b9a8-ad8a-40e3-955c-53aed07a9140" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.294575 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.298352 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" event={"ID":"88c9b59f-1809-4252-a8fc-cad965848dc0","Type":"ContainerStarted","Data":"647380be5685fa6b50f7f792740b85f949e8f14590a8e512665afee4f0b8f232"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.304323 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" event={"ID":"8d7ccd8e-549f-4b23-bcec-4a1c14e10478","Type":"ContainerStarted","Data":"8abe32a851225fd5d1922fb4bae72a41eee4fece3bf7728f299ea9031439b8ed"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.304365 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" event={"ID":"8d7ccd8e-549f-4b23-bcec-4a1c14e10478","Type":"ContainerStarted","Data":"3eb03704800aef4b382d689fe70eb2381222358dad2e910378a0a21f7f51d89c"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.314451 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:47 crc kubenswrapper[4874]: E0122 11:42:47.314840 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:47.814825139 +0000 UTC m=+141.659896209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.326562 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" event={"ID":"67454032-f4d2-418e-afad-a48c0a80007d","Type":"ContainerStarted","Data":"df3978feba47cd340cd15446fb8141642ed468d9027e1ea9b0efb6ffb5809622"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.326679 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" event={"ID":"67454032-f4d2-418e-afad-a48c0a80007d","Type":"ContainerStarted","Data":"38bfbedd5752105bf5782fa31806e1ec9b484b61fdf5b89b62091b2355d9e701"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.330913 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" event={"ID":"4f61e177-ea8e-4f61-8ba2-67906f08a00c","Type":"ContainerStarted","Data":"558b2d28a2ab0e3973110f8b8a82c613077883521bfb47a8a2b98b587b2161ce"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.334926 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" event={"ID":"0a15050a-7cbd-40f6-a656-a68293c0878a","Type":"ContainerStarted","Data":"059406f38b7eefe033761eb54f07a9cd0f802e07f88ff68ccbdb582fae9d8d28"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.334963 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" event={"ID":"0a15050a-7cbd-40f6-a656-a68293c0878a","Type":"ContainerStarted","Data":"c99ae11d334a79b6695ceefd2cce7368234bf383f78e3abddc229aa3df93fe78"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.337422 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" event={"ID":"51b2c452-70bd-4a3e-bd74-c7cd6513fd45","Type":"ContainerStarted","Data":"cd1de5446be5dd102373100e35d09ee8ab1132fadaf45b1ae27a4679de5f15b0"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.341272 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" event={"ID":"8b5afeda-c180-4aa9-a831-a0840d495fa8","Type":"ContainerStarted","Data":"a8a6f08c614f9afd8df854c6dbc31398112a667166fd12bd4926f9e1ab58cbc0"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.341305 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" event={"ID":"8b5afeda-c180-4aa9-a831-a0840d495fa8","Type":"ContainerStarted","Data":"98a3c21d4aff6ce8285894f5ee62debaac34c5731c09beac5a15c46753df56eb"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.342077 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.347769 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" event={"ID":"65f71c2e-ab34-4d33-905f-609555dab78c","Type":"ContainerStarted","Data":"84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.348142 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" event={"ID":"65f71c2e-ab34-4d33-905f-609555dab78c","Type":"ContainerStarted","Data":"a9c651399a50d57c614df95905597a481669874ba9ca8205cf5d6198908fbe7c"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.348169 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.351087 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" event={"ID":"f745af09-93ed-4c94-8411-f14b0aaaf1cf","Type":"ContainerStarted","Data":"e0c8fda880a7d1c41dcfeb90cb150869d269ef1f7a3b5eb0bae412e64024a346"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.351124 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" event={"ID":"f745af09-93ed-4c94-8411-f14b0aaaf1cf","Type":"ContainerStarted","Data":"acd01a69492923986ce95bf35e608599514c2e340da10f8fc96437bf69c498a4"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.352773 4874 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-dt8w6 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.352817 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" podUID="8b5afeda-c180-4aa9-a831-a0840d495fa8" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.353068 4874 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-2bn87 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.353105 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" podUID="65f71c2e-ab34-4d33-905f-609555dab78c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.360335 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kxvzk" event={"ID":"41a97907-0515-44cd-b6a4-87463a9b3819","Type":"ContainerStarted","Data":"64fc50e2df0fd1b74488e28f23047834bb85e5066ec5debf1bf297a69177c9bd"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.370984 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" event={"ID":"4ee5bb84-2932-4f1c-8af7-39e10c1260c2","Type":"ContainerStarted","Data":"fdc414e4a722957272fbe2e844434bbeab67b1a4d2e7f67709f7f840c0fadd59"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.374807 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" event={"ID":"a3d1224e-a43e-4dca-8ad0-239dc50b6d58","Type":"ContainerStarted","Data":"99e494f37f7ff1730b693230e1cbf965963ea9b1694d43f3655c495b87182e96"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.374854 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" event={"ID":"a3d1224e-a43e-4dca-8ad0-239dc50b6d58","Type":"ContainerStarted","Data":"1abeec4079be20567280eedb982b90bb26126680ec7f316575371854a41abc94"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.415642 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s5xxd" event={"ID":"a54b65f9-405b-4a90-a9eb-e0438722ffa8","Type":"ContainerStarted","Data":"e053f4fe36e9cf751669f7b77ee2f81f3c10d8d96e7ee32d01620e32539d1ee6"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.415682 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:47 crc kubenswrapper[4874]: E0122 11:42:47.415929 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:47.915917547 +0000 UTC m=+141.760988607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.438606 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vpjnl" podStartSLOduration=123.438579351 podStartE2EDuration="2m3.438579351s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:47.436197466 +0000 UTC m=+141.281268546" watchObservedRunningTime="2026-01-22 11:42:47.438579351 +0000 UTC m=+141.283650421" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.440603 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" event={"ID":"4370babd-21ea-4c7e-81b9-cdd611127094","Type":"ContainerStarted","Data":"e9d4fd6d2d34a9ae6171b6295d157abe4e467646bf8eab5eaeda4c6c84d3e4ea"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.456708 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" event={"ID":"574225b6-a476-4591-9be9-ddd94e5281ef","Type":"ContainerStarted","Data":"8f66be03fbd08062b606ad0121719d2dd562e0fa71df10028a45ad92e295ecf2"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.456750 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" event={"ID":"574225b6-a476-4591-9be9-ddd94e5281ef","Type":"ContainerStarted","Data":"7b30c3eda188977dd3f3fc4a6dd413c05a8e5c43f35ad2de13267b93fe586d76"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.457635 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.468610 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" event={"ID":"288e6358-c74b-4597-8968-726a31365f82","Type":"ContainerStarted","Data":"6bd7be9aa27d666be381a7fd11c421fab5dcf5107ce195055ac7e95d943119f5"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.468780 4874 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-b6c5m container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.468835 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" podUID="574225b6-a476-4591-9be9-ddd94e5281ef" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.473934 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" event={"ID":"f87cb481-5261-4028-99e2-cd57ff6b61e1","Type":"ContainerStarted","Data":"ee99f08d20b4163a0b446d289abf922985a606f83627f0a4c1a459ac62475a83"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.486820 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" event={"ID":"213d34f5-75cd-459c-9e56-2938fe5e3950","Type":"ContainerStarted","Data":"a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.487303 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.494975 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" podStartSLOduration=122.494961998 podStartE2EDuration="2m2.494961998s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:47.469499736 +0000 UTC m=+141.314570806" watchObservedRunningTime="2026-01-22 11:42:47.494961998 +0000 UTC m=+141.340033058" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.503554 4874 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-6dvvn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.503622 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" podUID="213d34f5-75cd-459c-9e56-2938fe5e3950" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.519325 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.519699 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6lgxn" podStartSLOduration=122.519685759 podStartE2EDuration="2m2.519685759s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:47.493857564 +0000 UTC m=+141.338928634" watchObservedRunningTime="2026-01-22 11:42:47.519685759 +0000 UTC m=+141.364756839" Jan 22 11:42:47 crc kubenswrapper[4874]: E0122 11:42:47.520496 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.020485013 +0000 UTC m=+141.865556083 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.521751 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fmhx8" podStartSLOduration=123.521740063 podStartE2EDuration="2m3.521740063s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:47.518812531 +0000 UTC m=+141.363883601" watchObservedRunningTime="2026-01-22 11:42:47.521740063 +0000 UTC m=+141.366811133" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.548252 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk" event={"ID":"a785e748-a557-4c24-8a4d-dc03cfc3c357","Type":"ContainerStarted","Data":"8a7e18cdf85973b3ea4e51ff8b10e08605086cbc0498e73cb8e3eb57f83837e6"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.548505 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk" event={"ID":"a785e748-a557-4c24-8a4d-dc03cfc3c357","Type":"ContainerStarted","Data":"b7ee0f5f651238d901bddc9fc4a5a84a9fa512ea93cb44bd4d6c58d7ba62b5ae"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.581598 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vsxx4" podStartSLOduration=123.581578199 podStartE2EDuration="2m3.581578199s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:47.543037384 +0000 UTC m=+141.388108454" watchObservedRunningTime="2026-01-22 11:42:47.581578199 +0000 UTC m=+141.426649269" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.581995 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6zpzc" podStartSLOduration=122.581991283 podStartE2EDuration="2m2.581991283s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:47.580481095 +0000 UTC m=+141.425552165" watchObservedRunningTime="2026-01-22 11:42:47.581991283 +0000 UTC m=+141.427062353" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.593244 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" event={"ID":"1795c220-db74-434e-9111-917ff6d95077","Type":"ContainerStarted","Data":"afefa41de116e68199e922f05b1e872d5782f38e366fdcfa816791519a8ca510"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.593433 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" event={"ID":"1795c220-db74-434e-9111-917ff6d95077","Type":"ContainerStarted","Data":"e17af57e0f59e83ba770c3db737189f5369bc87bbd276ea5eb99f256b0202cd7"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.607660 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" event={"ID":"6a39a4e6-a146-4ff3-8e90-eb516fe27e38","Type":"ContainerStarted","Data":"295b2c7f753febed4403ab96ad5c95131d05455fa1f58fbc53b3147f90ba7b5b"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.607697 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" event={"ID":"6a39a4e6-a146-4ff3-8e90-eb516fe27e38","Type":"ContainerStarted","Data":"0be9d83f8aee7488de8a04bf682b44f0aa9194a9f62592e763949214d15c368b"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.614831 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" podStartSLOduration=122.614814528 podStartE2EDuration="2m2.614814528s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:47.613553218 +0000 UTC m=+141.458624278" watchObservedRunningTime="2026-01-22 11:42:47.614814528 +0000 UTC m=+141.459885598" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.623724 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:47 crc kubenswrapper[4874]: E0122 11:42:47.626204 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.126187516 +0000 UTC m=+141.971258586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.631792 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l87tw" event={"ID":"7a9c3919-d376-4655-818e-3a88a8c0b883","Type":"ContainerStarted","Data":"d814de8e63d732aeae1c15bab08a600f371f9751ea8d100cd5ddd91e02fd32ed"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.654793 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" podStartSLOduration=122.654774847 podStartE2EDuration="2m2.654774847s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:47.651735362 +0000 UTC m=+141.496806422" watchObservedRunningTime="2026-01-22 11:42:47.654774847 +0000 UTC m=+141.499845917" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.666503 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s4fb2" event={"ID":"d80367da-a8d4-4b76-9b33-5526aad85229","Type":"ContainerStarted","Data":"416bed418a7c55213c0f7dd0edc12db3eb1e52d960bc08ec7924166a1b9043bc"} Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.677000 4874 patch_prober.go:28] interesting pod/downloads-7954f5f757-jgw4c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.677054 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jgw4c" podUID="3b9983ba-ed8d-4654-ba74-f25433aa7ee7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.691149 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.697610 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fl8nd" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.715174 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-mftbv" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.747611 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:47 crc kubenswrapper[4874]: E0122 11:42:47.761993 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.261967767 +0000 UTC m=+142.107038837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.822920 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" podStartSLOduration=122.822895438 podStartE2EDuration="2m2.822895438s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:47.822520526 +0000 UTC m=+141.667591606" watchObservedRunningTime="2026-01-22 11:42:47.822895438 +0000 UTC m=+141.667966508" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.859901 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:47 crc kubenswrapper[4874]: E0122 11:42:47.866091 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.36607274 +0000 UTC m=+142.211143810 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.941340 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-s5xxd" podStartSLOduration=7.941315682 podStartE2EDuration="7.941315682s" podCreationTimestamp="2026-01-22 11:42:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:47.938628438 +0000 UTC m=+141.783699508" watchObservedRunningTime="2026-01-22 11:42:47.941315682 +0000 UTC m=+141.786386762" Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.963537 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:47 crc kubenswrapper[4874]: E0122 11:42:47.963868 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.463846433 +0000 UTC m=+142.308917503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:47 crc kubenswrapper[4874]: I0122 11:42:47.963940 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:47 crc kubenswrapper[4874]: E0122 11:42:47.964827 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.464806883 +0000 UTC m=+142.309877953 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.044821 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-dwcjx" podStartSLOduration=123.044803365 podStartE2EDuration="2m3.044803365s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:48.042001717 +0000 UTC m=+141.887072787" watchObservedRunningTime="2026-01-22 11:42:48.044803365 +0000 UTC m=+141.889874435" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.064738 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.065068 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.565051784 +0000 UTC m=+142.410122854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.145146 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6jqjk" podStartSLOduration=123.145132308 podStartE2EDuration="2m3.145132308s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:48.144350304 +0000 UTC m=+141.989421374" watchObservedRunningTime="2026-01-22 11:42:48.145132308 +0000 UTC m=+141.990203378" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.145788 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-sjhjk" podStartSLOduration=124.145783559 podStartE2EDuration="2m4.145783559s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:48.113599264 +0000 UTC m=+141.958670344" watchObservedRunningTime="2026-01-22 11:42:48.145783559 +0000 UTC m=+141.990854629" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.166645 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.167126 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.667110461 +0000 UTC m=+142.512181531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.203609 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" podStartSLOduration=123.203589481 podStartE2EDuration="2m3.203589481s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:48.19274612 +0000 UTC m=+142.037817190" watchObservedRunningTime="2026-01-22 11:42:48.203589481 +0000 UTC m=+142.048660551" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.244626 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:48 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:48 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:48 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.245212 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.267709 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.268095 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.768080755 +0000 UTC m=+142.613151825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.369544 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.369966 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.869955147 +0000 UTC m=+142.715026217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.471001 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.471314 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.971283412 +0000 UTC m=+142.816354482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.471608 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.472036 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:48.972021185 +0000 UTC m=+142.817092255 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.572993 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.573418 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.073376971 +0000 UTC m=+142.918448051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.672327 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" event={"ID":"4f61e177-ea8e-4f61-8ba2-67906f08a00c","Type":"ContainerStarted","Data":"e2ad4ec597599d88f96fd683a147efd30f859f39b2241cce51d81aefd182a0d0"} Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.672371 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" event={"ID":"4f61e177-ea8e-4f61-8ba2-67906f08a00c","Type":"ContainerStarted","Data":"edd0852932e3487ad47d142c96cedcef6f74c46c481681268729bdeff6368ae4"} Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.674567 4874 generic.go:334] "Generic (PLEG): container finished" podID="288e6358-c74b-4597-8968-726a31365f82" containerID="e768dbe1f074a94bc522533a48c9c56cdc2451138e0425b58698e0fefde2a767" exitCode=0 Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.674653 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" event={"ID":"288e6358-c74b-4597-8968-726a31365f82","Type":"ContainerDied","Data":"e768dbe1f074a94bc522533a48c9c56cdc2451138e0425b58698e0fefde2a767"} Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.676034 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.676415 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.176383209 +0000 UTC m=+143.021454279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.676592 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" event={"ID":"f87cb481-5261-4028-99e2-cd57ff6b61e1","Type":"ContainerStarted","Data":"db2bcd42d1984929d94787587108e648f04aeadd797d990928b45e142f5e0fed"} Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.682778 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kxvzk" event={"ID":"41a97907-0515-44cd-b6a4-87463a9b3819","Type":"ContainerStarted","Data":"a5103129246337de5be9e1e278501f97850bb06e8547497811d03ce40d6aebca"} Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.682820 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kxvzk" event={"ID":"41a97907-0515-44cd-b6a4-87463a9b3819","Type":"ContainerStarted","Data":"969bbedcf0c44414a0fb98fdd634b6bec509293125e4834ca9edd7247b104f5a"} Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.682987 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.688796 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" event={"ID":"1795c220-db74-434e-9111-917ff6d95077","Type":"ContainerStarted","Data":"9e86d6331e2e5bfe460602c5aec3829a0d20d45fbf3f2f0800246076bb46d1fb"} Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.693241 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l87tw" event={"ID":"7a9c3919-d376-4655-818e-3a88a8c0b883","Type":"ContainerStarted","Data":"517858dad1bddae80dce4e4109465357192315c04012a6485f483b33bc958c8a"} Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.698300 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-rx6nv" podStartSLOduration=124.69828334 podStartE2EDuration="2m4.69828334s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:48.695916274 +0000 UTC m=+142.540987334" watchObservedRunningTime="2026-01-22 11:42:48.69828334 +0000 UTC m=+142.543354420" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.703654 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" event={"ID":"8d7ccd8e-549f-4b23-bcec-4a1c14e10478","Type":"ContainerStarted","Data":"99b89d43849311e9d542317356dd7454d7f1912dc8012e16fc9ff5ad2f45234b"} Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.703907 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.709675 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" event={"ID":"67454032-f4d2-418e-afad-a48c0a80007d","Type":"ContainerStarted","Data":"fcf228c9117a38a0da1eb2447ee8dd31ed6ace0a705bd6cf118df239b6c7105e"} Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.710962 4874 patch_prober.go:28] interesting pod/downloads-7954f5f757-jgw4c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.711011 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jgw4c" podUID="3b9983ba-ed8d-4654-ba74-f25433aa7ee7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.711708 4874 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-6dvvn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.711754 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" podUID="213d34f5-75cd-459c-9e56-2938fe5e3950" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.732039 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-f64ps" podStartSLOduration=123.732021143 podStartE2EDuration="2m3.732021143s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:48.730574428 +0000 UTC m=+142.575645508" watchObservedRunningTime="2026-01-22 11:42:48.732021143 +0000 UTC m=+142.577092213" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.750061 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.750099 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dt8w6" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.750130 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b6c5m" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.776951 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.777188 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.277130245 +0000 UTC m=+143.122201315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.777451 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.778784 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.278775307 +0000 UTC m=+143.123846377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.807173 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gsfnm" podStartSLOduration=124.807151212 podStartE2EDuration="2m4.807151212s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:48.806864513 +0000 UTC m=+142.651935603" watchObservedRunningTime="2026-01-22 11:42:48.807151212 +0000 UTC m=+142.652222282" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.840548 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-kxvzk" podStartSLOduration=8.840526195 podStartE2EDuration="8.840526195s" podCreationTimestamp="2026-01-22 11:42:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:48.834103392 +0000 UTC m=+142.679174462" watchObservedRunningTime="2026-01-22 11:42:48.840526195 +0000 UTC m=+142.685597275" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.880597 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gv9x5" Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.882157 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.887815 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.387774714 +0000 UTC m=+143.232845784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.988184 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:48 crc kubenswrapper[4874]: E0122 11:42:48.989085 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.489072168 +0000 UTC m=+143.334143238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:48 crc kubenswrapper[4874]: I0122 11:42:48.990703 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-89ldw" podStartSLOduration=124.990674068 podStartE2EDuration="2m4.990674068s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:48.987358094 +0000 UTC m=+142.832429164" watchObservedRunningTime="2026-01-22 11:42:48.990674068 +0000 UTC m=+142.835745148" Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.017871 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" podStartSLOduration=124.017847545 podStartE2EDuration="2m4.017847545s" podCreationTimestamp="2026-01-22 11:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:49.014709246 +0000 UTC m=+142.859780316" watchObservedRunningTime="2026-01-22 11:42:49.017847545 +0000 UTC m=+142.862918625" Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.089357 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.089573 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.589543996 +0000 UTC m=+143.434615066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.089815 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.090110 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.590099574 +0000 UTC m=+143.435170644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.190794 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.191377 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.691362106 +0000 UTC m=+143.536433176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.243835 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:49 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:49 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:49 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.243889 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.292923 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.293672 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.793655291 +0000 UTC m=+143.638726351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.394776 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.394921 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.894899074 +0000 UTC m=+143.739970154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.395161 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.395557 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.895546654 +0000 UTC m=+143.740617724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.496676 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.496861 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.996830277 +0000 UTC m=+143.841901347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.496942 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.497524 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:49.997509269 +0000 UTC m=+143.842580339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.598039 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.598241 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:50.098212914 +0000 UTC m=+143.943283984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.598529 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.598913 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:50.098899626 +0000 UTC m=+143.943970696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.661475 4874 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.699497 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.699826 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:50.199810828 +0000 UTC m=+144.044881898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.714700 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l87tw" event={"ID":"7a9c3919-d376-4655-818e-3a88a8c0b883","Type":"ContainerStarted","Data":"4a38d036b900efde16ecd783e642919b32ccb03f0034722b91a174d22da7d67e"} Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.714745 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l87tw" event={"ID":"7a9c3919-d376-4655-818e-3a88a8c0b883","Type":"ContainerStarted","Data":"0cd6eec0f86dfdb5ce52d515d3ada676af7aaac76c0a55e70afa6045c1ecc6b0"} Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.719167 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" event={"ID":"288e6358-c74b-4597-8968-726a31365f82","Type":"ContainerStarted","Data":"1fe1efdd8b19b54cbe4b074c9d510c5de89756de0dabe35ff3a33d6d3b8c34db"} Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.719198 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" event={"ID":"288e6358-c74b-4597-8968-726a31365f82","Type":"ContainerStarted","Data":"0382a03db8e05c46b9dbc58b391e1572cd8dd298eda6837a0b9c11a6730a974b"} Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.741463 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" podStartSLOduration=125.74144348 podStartE2EDuration="2m5.74144348s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:49.740528821 +0000 UTC m=+143.585599891" watchObservedRunningTime="2026-01-22 11:42:49.74144348 +0000 UTC m=+143.586514550" Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.802036 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.803497 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 11:42:50.303478586 +0000 UTC m=+144.148549656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jg4wj" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.816037 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bk5qd"] Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.817058 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.820892 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.828708 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bk5qd"] Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.890583 4874 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-22T11:42:49.661712986Z","Handler":null,"Name":""} Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.903626 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.903893 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9tqj\" (UniqueName: \"kubernetes.io/projected/829e346d-eb89-4705-83c4-99d02fca8971-kube-api-access-k9tqj\") pod \"community-operators-bk5qd\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.903925 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-catalog-content\") pod \"community-operators-bk5qd\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.903955 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-utilities\") pod \"community-operators-bk5qd\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:42:49 crc kubenswrapper[4874]: E0122 11:42:49.904118 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 11:42:50.404090018 +0000 UTC m=+144.249161088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.923434 4874 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 22 11:42:49 crc kubenswrapper[4874]: I0122 11:42:49.923665 4874 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.003512 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r2mwt"] Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.004760 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.005154 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-catalog-content\") pod \"community-operators-bk5qd\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.005264 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-utilities\") pod \"community-operators-bk5qd\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.005409 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.005555 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9tqj\" (UniqueName: \"kubernetes.io/projected/829e346d-eb89-4705-83c4-99d02fca8971-kube-api-access-k9tqj\") pod \"community-operators-bk5qd\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.006293 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-utilities\") pod \"community-operators-bk5qd\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.006357 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-catalog-content\") pod \"community-operators-bk5qd\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.006799 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.009114 4874 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.009145 4874 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.016116 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r2mwt"] Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.023775 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9tqj\" (UniqueName: \"kubernetes.io/projected/829e346d-eb89-4705-83c4-99d02fca8971-kube-api-access-k9tqj\") pod \"community-operators-bk5qd\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.035424 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jg4wj\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.107158 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.107387 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2xvq\" (UniqueName: \"kubernetes.io/projected/dea4a6eb-c0b1-432a-81f2-e417250b0138-kube-api-access-w2xvq\") pod \"certified-operators-r2mwt\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.107428 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-catalog-content\") pod \"certified-operators-r2mwt\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.107444 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-utilities\") pod \"certified-operators-r2mwt\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.119983 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.154784 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.205905 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k7sj5"] Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.207073 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.208213 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2xvq\" (UniqueName: \"kubernetes.io/projected/dea4a6eb-c0b1-432a-81f2-e417250b0138-kube-api-access-w2xvq\") pod \"certified-operators-r2mwt\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.208254 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-catalog-content\") pod \"certified-operators-r2mwt\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.208277 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-utilities\") pod \"certified-operators-r2mwt\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.208774 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-utilities\") pod \"certified-operators-r2mwt\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.209349 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-catalog-content\") pod \"certified-operators-r2mwt\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.219500 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k7sj5"] Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.227558 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2xvq\" (UniqueName: \"kubernetes.io/projected/dea4a6eb-c0b1-432a-81f2-e417250b0138-kube-api-access-w2xvq\") pod \"certified-operators-r2mwt\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.235847 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.240711 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:50 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:50 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:50 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.240751 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.310470 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-catalog-content\") pod \"community-operators-k7sj5\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.310769 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djwcc\" (UniqueName: \"kubernetes.io/projected/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-kube-api-access-djwcc\") pod \"community-operators-k7sj5\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.310845 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-utilities\") pod \"community-operators-k7sj5\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.317249 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.411178 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cxrx9"] Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.411860 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-catalog-content\") pod \"community-operators-k7sj5\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.411925 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djwcc\" (UniqueName: \"kubernetes.io/projected/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-kube-api-access-djwcc\") pod \"community-operators-k7sj5\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.412028 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-utilities\") pod \"community-operators-k7sj5\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.413213 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-catalog-content\") pod \"community-operators-k7sj5\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.413886 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-utilities\") pod \"community-operators-k7sj5\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.414390 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.419216 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cxrx9"] Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.431942 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djwcc\" (UniqueName: \"kubernetes.io/projected/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-kube-api-access-djwcc\") pod \"community-operators-k7sj5\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.494070 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bk5qd"] Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.514322 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzmfp\" (UniqueName: \"kubernetes.io/projected/ede22568-7d51-42d6-b96d-1783c4e5b370-kube-api-access-wzmfp\") pod \"certified-operators-cxrx9\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.514376 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-utilities\") pod \"certified-operators-cxrx9\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.514441 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-catalog-content\") pod \"certified-operators-cxrx9\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.531560 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:42:50 crc kubenswrapper[4874]: W0122 11:42:50.532649 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod829e346d_eb89_4705_83c4_99d02fca8971.slice/crio-06e653a8a66624e58a13fce521592f82bf5da1e8eff383092cdaf8dbc041daca WatchSource:0}: Error finding container 06e653a8a66624e58a13fce521592f82bf5da1e8eff383092cdaf8dbc041daca: Status 404 returned error can't find the container with id 06e653a8a66624e58a13fce521592f82bf5da1e8eff383092cdaf8dbc041daca Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.535796 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jg4wj"] Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.615179 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-catalog-content\") pod \"certified-operators-cxrx9\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.615286 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzmfp\" (UniqueName: \"kubernetes.io/projected/ede22568-7d51-42d6-b96d-1783c4e5b370-kube-api-access-wzmfp\") pod \"certified-operators-cxrx9\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.615306 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-utilities\") pod \"certified-operators-cxrx9\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.615972 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-utilities\") pod \"certified-operators-cxrx9\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.619673 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-catalog-content\") pod \"certified-operators-cxrx9\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.628678 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r2mwt"] Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.640361 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzmfp\" (UniqueName: \"kubernetes.io/projected/ede22568-7d51-42d6-b96d-1783c4e5b370-kube-api-access-wzmfp\") pod \"certified-operators-cxrx9\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.744766 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:42:50 crc kubenswrapper[4874]: I0122 11:42:50.772886 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-l87tw" podStartSLOduration=10.77286776 podStartE2EDuration="10.77286776s" podCreationTimestamp="2026-01-22 11:42:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:50.754896254 +0000 UTC m=+144.599967324" watchObservedRunningTime="2026-01-22 11:42:50.77286776 +0000 UTC m=+144.617938830" Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.043418 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.044180 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bk5qd" event={"ID":"829e346d-eb89-4705-83c4-99d02fca8971","Type":"ContainerStarted","Data":"06e653a8a66624e58a13fce521592f82bf5da1e8eff383092cdaf8dbc041daca"} Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.044208 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2mwt" event={"ID":"dea4a6eb-c0b1-432a-81f2-e417250b0138","Type":"ContainerStarted","Data":"44f0f27f3f6817f2e87df4da350d3341a31946833627d7893617f4278baf12f8"} Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.044220 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k7sj5"] Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.044233 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" event={"ID":"90a037cd-cacf-4706-a857-f65c8f16c384","Type":"ContainerStarted","Data":"61abb1ca8d94b8a93a2470a92da41ddacd13e2ac8b6f3b50b461fb0f48f44b61"} Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.044243 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l87tw" event={"ID":"7a9c3919-d376-4655-818e-3a88a8c0b883","Type":"ContainerStarted","Data":"cb0e942e428230d003297b16951490d42d8fa2b63dde46bfa4bdd15dc61317fc"} Jan 22 11:42:51 crc kubenswrapper[4874]: W0122 11:42:51.050133 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d6b2fd5_d040_4be5_afb1_a375bddd3e88.slice/crio-d963c7b2a3b3c57b49606220d41e2548892102c0a65c18048da0d0bded52f8de WatchSource:0}: Error finding container d963c7b2a3b3c57b49606220d41e2548892102c0a65c18048da0d0bded52f8de: Status 404 returned error can't find the container with id d963c7b2a3b3c57b49606220d41e2548892102c0a65c18048da0d0bded52f8de Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.236224 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:51 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:51 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:51 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.236564 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.365316 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cxrx9"] Jan 22 11:42:51 crc kubenswrapper[4874]: W0122 11:42:51.421808 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podede22568_7d51_42d6_b96d_1783c4e5b370.slice/crio-eb9fe6572677e18214bcf1bd4f031854d94ab0f4aacf889a0900f36d0c7491fc WatchSource:0}: Error finding container eb9fe6572677e18214bcf1bd4f031854d94ab0f4aacf889a0900f36d0c7491fc: Status 404 returned error can't find the container with id eb9fe6572677e18214bcf1bd4f031854d94ab0f4aacf889a0900f36d0c7491fc Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.734483 4874 generic.go:334] "Generic (PLEG): container finished" podID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" containerID="ab54d9fab26a3ef6698d354ee6813928ad7aedd283867428b79b6242016d11de" exitCode=0 Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.734801 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7sj5" event={"ID":"7d6b2fd5-d040-4be5-afb1-a375bddd3e88","Type":"ContainerDied","Data":"ab54d9fab26a3ef6698d354ee6813928ad7aedd283867428b79b6242016d11de"} Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.734827 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7sj5" event={"ID":"7d6b2fd5-d040-4be5-afb1-a375bddd3e88","Type":"ContainerStarted","Data":"d963c7b2a3b3c57b49606220d41e2548892102c0a65c18048da0d0bded52f8de"} Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.736326 4874 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.737128 4874 generic.go:334] "Generic (PLEG): container finished" podID="ede22568-7d51-42d6-b96d-1783c4e5b370" containerID="769e221159e673ec03f3822d5d4059809f4101a42a4ddda5d439bdab06ed4064" exitCode=0 Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.737265 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxrx9" event={"ID":"ede22568-7d51-42d6-b96d-1783c4e5b370","Type":"ContainerDied","Data":"769e221159e673ec03f3822d5d4059809f4101a42a4ddda5d439bdab06ed4064"} Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.737333 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxrx9" event={"ID":"ede22568-7d51-42d6-b96d-1783c4e5b370","Type":"ContainerStarted","Data":"eb9fe6572677e18214bcf1bd4f031854d94ab0f4aacf889a0900f36d0c7491fc"} Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.739370 4874 generic.go:334] "Generic (PLEG): container finished" podID="829e346d-eb89-4705-83c4-99d02fca8971" containerID="e04166d257b1ab6ec62d2fb4997b1ee33b374447294ba53534e3f997572dbcbc" exitCode=0 Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.739442 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bk5qd" event={"ID":"829e346d-eb89-4705-83c4-99d02fca8971","Type":"ContainerDied","Data":"e04166d257b1ab6ec62d2fb4997b1ee33b374447294ba53534e3f997572dbcbc"} Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.740971 4874 generic.go:334] "Generic (PLEG): container finished" podID="dea4a6eb-c0b1-432a-81f2-e417250b0138" containerID="f102fff60f0bad6a90aabfcc1da36f807499c6f5b663e355666a3b9418698f61" exitCode=0 Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.741005 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2mwt" event={"ID":"dea4a6eb-c0b1-432a-81f2-e417250b0138","Type":"ContainerDied","Data":"f102fff60f0bad6a90aabfcc1da36f807499c6f5b663e355666a3b9418698f61"} Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.743673 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" event={"ID":"90a037cd-cacf-4706-a857-f65c8f16c384","Type":"ContainerStarted","Data":"e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe"} Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.796997 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" podStartSLOduration=127.796967691 podStartE2EDuration="2m7.796967691s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:42:51.79347108 +0000 UTC m=+145.638542170" watchObservedRunningTime="2026-01-22 11:42:51.796967691 +0000 UTC m=+145.642038771" Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.805962 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xhb8k"] Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.806965 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.808546 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.839198 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhb8k"] Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.895260 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.895881 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.897689 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.897874 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.902896 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.943879 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-catalog-content\") pod \"redhat-marketplace-xhb8k\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.943937 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-utilities\") pod \"redhat-marketplace-xhb8k\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:42:51 crc kubenswrapper[4874]: I0122 11:42:51.944065 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79nw7\" (UniqueName: \"kubernetes.io/projected/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-kube-api-access-79nw7\") pod \"redhat-marketplace-xhb8k\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.045650 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5bfa0c46-abaf-4d63-964c-d8a91fafbc82\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.045720 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-catalog-content\") pod \"redhat-marketplace-xhb8k\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.045833 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-utilities\") pod \"redhat-marketplace-xhb8k\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.045952 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5bfa0c46-abaf-4d63-964c-d8a91fafbc82\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.046133 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79nw7\" (UniqueName: \"kubernetes.io/projected/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-kube-api-access-79nw7\") pod \"redhat-marketplace-xhb8k\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.046292 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-catalog-content\") pod \"redhat-marketplace-xhb8k\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.046306 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-utilities\") pod \"redhat-marketplace-xhb8k\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.064218 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79nw7\" (UniqueName: \"kubernetes.io/projected/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-kube-api-access-79nw7\") pod \"redhat-marketplace-xhb8k\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.123174 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.147222 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5bfa0c46-abaf-4d63-964c-d8a91fafbc82\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.147311 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5bfa0c46-abaf-4d63-964c-d8a91fafbc82\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.147373 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5bfa0c46-abaf-4d63-964c-d8a91fafbc82\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.161972 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5bfa0c46-abaf-4d63-964c-d8a91fafbc82\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.212268 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pbbps"] Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.213855 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.221656 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.226364 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbbps"] Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.241289 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:52 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:52 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:52 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.241331 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.351100 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-utilities\") pod \"redhat-marketplace-pbbps\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.351198 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-catalog-content\") pod \"redhat-marketplace-pbbps\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.351316 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrxzq\" (UniqueName: \"kubernetes.io/projected/194d73ac-fe2b-4e80-b03b-c1b780b55990-kube-api-access-jrxzq\") pod \"redhat-marketplace-pbbps\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.374525 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhb8k"] Jan 22 11:42:52 crc kubenswrapper[4874]: W0122 11:42:52.422463 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09a47d07_9bf6_4033_8c08_cc3aef9fe4f4.slice/crio-4f0dfc138b95fd589daf732c95563b35b123c0075950db736b607653c673e3e1 WatchSource:0}: Error finding container 4f0dfc138b95fd589daf732c95563b35b123c0075950db736b607653c673e3e1: Status 404 returned error can't find the container with id 4f0dfc138b95fd589daf732c95563b35b123c0075950db736b607653c673e3e1 Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.452966 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-catalog-content\") pod \"redhat-marketplace-pbbps\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.453065 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrxzq\" (UniqueName: \"kubernetes.io/projected/194d73ac-fe2b-4e80-b03b-c1b780b55990-kube-api-access-jrxzq\") pod \"redhat-marketplace-pbbps\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.453109 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-utilities\") pod \"redhat-marketplace-pbbps\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.453544 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-catalog-content\") pod \"redhat-marketplace-pbbps\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.453555 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-utilities\") pod \"redhat-marketplace-pbbps\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.471707 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrxzq\" (UniqueName: \"kubernetes.io/projected/194d73ac-fe2b-4e80-b03b-c1b780b55990-kube-api-access-jrxzq\") pod \"redhat-marketplace-pbbps\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.495569 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.539199 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.560741 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.757309 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.757344 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.757391 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.765213 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.781467 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.790138 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.820320 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5bfa0c46-abaf-4d63-964c-d8a91fafbc82","Type":"ContainerStarted","Data":"38b9f2be77edd43a5d803726e5294fbc3482629ac4f147f216e1e74861fc0fee"} Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.838732 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.848626 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.860047 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.867087 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.890177 4874 generic.go:334] "Generic (PLEG): container finished" podID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerID="af7c8535e94f5d42ca318e78dacd324021ec14636cac5b8d6617938e7cede85c" exitCode=0 Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.891544 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhb8k" event={"ID":"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4","Type":"ContainerDied","Data":"af7c8535e94f5d42ca318e78dacd324021ec14636cac5b8d6617938e7cede85c"} Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.891587 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhb8k" event={"ID":"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4","Type":"ContainerStarted","Data":"4f0dfc138b95fd589daf732c95563b35b123c0075950db736b607653c673e3e1"} Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.891638 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:42:52 crc kubenswrapper[4874]: I0122 11:42:52.904643 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbbps"] Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.013318 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s2gqx"] Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.016899 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.019368 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.026060 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s2gqx"] Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.107274 4874 patch_prober.go:28] interesting pod/downloads-7954f5f757-jgw4c container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.107328 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-jgw4c" podUID="3b9983ba-ed8d-4654-ba74-f25433aa7ee7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.107547 4874 patch_prober.go:28] interesting pod/downloads-7954f5f757-jgw4c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.107576 4874 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jgw4c" podUID="3b9983ba-ed8d-4654-ba74-f25433aa7ee7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.140675 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.163157 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-catalog-content\") pod \"redhat-operators-s2gqx\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.163748 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92txx\" (UniqueName: \"kubernetes.io/projected/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-kube-api-access-92txx\") pod \"redhat-operators-s2gqx\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.163811 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-utilities\") pod \"redhat-operators-s2gqx\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.173183 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.174560 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.189196 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.189234 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.191037 4874 patch_prober.go:28] interesting pod/console-f9d7485db-wws2s container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.191103 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wws2s" podUID="2dad3db6-cddd-457d-8efa-908257ef7cc5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.194550 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.235876 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.244971 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:53 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:53 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:53 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.245322 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.265904 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92txx\" (UniqueName: \"kubernetes.io/projected/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-kube-api-access-92txx\") pod \"redhat-operators-s2gqx\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.265952 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-utilities\") pod \"redhat-operators-s2gqx\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.265986 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-catalog-content\") pod \"redhat-operators-s2gqx\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.268226 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-utilities\") pod \"redhat-operators-s2gqx\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.269338 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-catalog-content\") pod \"redhat-operators-s2gqx\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.315481 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92txx\" (UniqueName: \"kubernetes.io/projected/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-kube-api-access-92txx\") pod \"redhat-operators-s2gqx\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.346737 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.431217 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-chgtv"] Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.440668 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.441090 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.463189 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-chgtv"] Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.575316 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v87lj\" (UniqueName: \"kubernetes.io/projected/25f55928-a244-44f3-83a5-b6bdf551bda6-kube-api-access-v87lj\") pod \"redhat-operators-chgtv\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.575369 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-catalog-content\") pod \"redhat-operators-chgtv\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.575535 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-utilities\") pod \"redhat-operators-chgtv\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.676620 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v87lj\" (UniqueName: \"kubernetes.io/projected/25f55928-a244-44f3-83a5-b6bdf551bda6-kube-api-access-v87lj\") pod \"redhat-operators-chgtv\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.676683 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-catalog-content\") pod \"redhat-operators-chgtv\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.676719 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-utilities\") pod \"redhat-operators-chgtv\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.677288 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-utilities\") pod \"redhat-operators-chgtv\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.677895 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-catalog-content\") pod \"redhat-operators-chgtv\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.697351 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v87lj\" (UniqueName: \"kubernetes.io/projected/25f55928-a244-44f3-83a5-b6bdf551bda6-kube-api-access-v87lj\") pod \"redhat-operators-chgtv\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.720844 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s2gqx"] Jan 22 11:42:53 crc kubenswrapper[4874]: W0122 11:42:53.757954 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d07947b_508d_4f12_ba1b_2d5f24a6db2c.slice/crio-880843b24ca67a3147842f7c7ec48125fab7fce5de031516ba04e79aff10f06a WatchSource:0}: Error finding container 880843b24ca67a3147842f7c7ec48125fab7fce5de031516ba04e79aff10f06a: Status 404 returned error can't find the container with id 880843b24ca67a3147842f7c7ec48125fab7fce5de031516ba04e79aff10f06a Jan 22 11:42:53 crc kubenswrapper[4874]: W0122 11:42:53.786764 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-8976c6b5a82cbf91c93238a03097b8c9df4789af32b7cf19a40426492a6a2a0b WatchSource:0}: Error finding container 8976c6b5a82cbf91c93238a03097b8c9df4789af32b7cf19a40426492a6a2a0b: Status 404 returned error can't find the container with id 8976c6b5a82cbf91c93238a03097b8c9df4789af32b7cf19a40426492a6a2a0b Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.826985 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.902369 4874 generic.go:334] "Generic (PLEG): container finished" podID="194d73ac-fe2b-4e80-b03b-c1b780b55990" containerID="1ecd914b99d6525c762a368e3ff1d7acae50f9b9d1238be939d00bb461aca834" exitCode=0 Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.902491 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbbps" event={"ID":"194d73ac-fe2b-4e80-b03b-c1b780b55990","Type":"ContainerDied","Data":"1ecd914b99d6525c762a368e3ff1d7acae50f9b9d1238be939d00bb461aca834"} Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.902519 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbbps" event={"ID":"194d73ac-fe2b-4e80-b03b-c1b780b55990","Type":"ContainerStarted","Data":"3b5a39abee6782e33316f7969db2ab23c5d711ceba05225f55e2ae3fb886847c"} Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.904238 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2gqx" event={"ID":"2d07947b-508d-4f12-ba1b-2d5f24a6db2c","Type":"ContainerStarted","Data":"880843b24ca67a3147842f7c7ec48125fab7fce5de031516ba04e79aff10f06a"} Jan 22 11:42:53 crc kubenswrapper[4874]: W0122 11:42:53.905280 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-86651d95f7a3ee7b3be1f4b424e626a1f03634119a3491f3dc562bc6709a4f14 WatchSource:0}: Error finding container 86651d95f7a3ee7b3be1f4b424e626a1f03634119a3491f3dc562bc6709a4f14: Status 404 returned error can't find the container with id 86651d95f7a3ee7b3be1f4b424e626a1f03634119a3491f3dc562bc6709a4f14 Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.910743 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7b1a3cd0f3714333cdad6aaa71b8709438bbefc59888541682aea1b2c0a1a6d2"} Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.910789 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d7338913815c5a3aedcf53c4f221dd3462e33965c1aac01ecf2ee057baef0eec"} Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.924297 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8976c6b5a82cbf91c93238a03097b8c9df4789af32b7cf19a40426492a6a2a0b"} Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.948064 4874 generic.go:334] "Generic (PLEG): container finished" podID="5bfa0c46-abaf-4d63-964c-d8a91fafbc82" containerID="127975d4720e9bff00fac314035fd53ba7989af2ade4f2461c7a1cbf032b68a0" exitCode=0 Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.948539 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5bfa0c46-abaf-4d63-964c-d8a91fafbc82","Type":"ContainerDied","Data":"127975d4720e9bff00fac314035fd53ba7989af2ade4f2461c7a1cbf032b68a0"} Jan 22 11:42:53 crc kubenswrapper[4874]: I0122 11:42:53.967618 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-cqkbr" Jan 22 11:42:54 crc kubenswrapper[4874]: I0122 11:42:54.235823 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:54 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:54 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:54 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:54 crc kubenswrapper[4874]: I0122 11:42:54.236044 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:54 crc kubenswrapper[4874]: I0122 11:42:54.322748 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-chgtv"] Jan 22 11:42:54 crc kubenswrapper[4874]: W0122 11:42:54.343542 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25f55928_a244_44f3_83a5_b6bdf551bda6.slice/crio-f695035bc56b57dcd6c7f8fc2e3f5c5198706ed0269a97d26b4dab2f6db84183 WatchSource:0}: Error finding container f695035bc56b57dcd6c7f8fc2e3f5c5198706ed0269a97d26b4dab2f6db84183: Status 404 returned error can't find the container with id f695035bc56b57dcd6c7f8fc2e3f5c5198706ed0269a97d26b4dab2f6db84183 Jan 22 11:42:54 crc kubenswrapper[4874]: I0122 11:42:54.990021 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1814b08775c033da9be344eecd2d9fdad98d56c9e03566ca118519e42a96b535"} Jan 22 11:42:54 crc kubenswrapper[4874]: I0122 11:42:54.990322 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:42:54 crc kubenswrapper[4874]: I0122 11:42:54.996567 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e7640c9b37b5d1dd90fc0a4bd9176f6aea51aec5b3333bc3882dc215a8ce61a4"} Jan 22 11:42:54 crc kubenswrapper[4874]: I0122 11:42:54.996627 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"86651d95f7a3ee7b3be1f4b424e626a1f03634119a3491f3dc562bc6709a4f14"} Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.038916 4874 generic.go:334] "Generic (PLEG): container finished" podID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerID="a94610599ca8c298d3b9c45ba5eaf696e08b7bd8dfa7298e41a50594597af12d" exitCode=0 Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.039218 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chgtv" event={"ID":"25f55928-a244-44f3-83a5-b6bdf551bda6","Type":"ContainerDied","Data":"a94610599ca8c298d3b9c45ba5eaf696e08b7bd8dfa7298e41a50594597af12d"} Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.039261 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chgtv" event={"ID":"25f55928-a244-44f3-83a5-b6bdf551bda6","Type":"ContainerStarted","Data":"f695035bc56b57dcd6c7f8fc2e3f5c5198706ed0269a97d26b4dab2f6db84183"} Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.052707 4874 generic.go:334] "Generic (PLEG): container finished" podID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerID="b45c0a0edb45e712987d45512f5f8181affde43a34f168d3078df992db0dd1ef" exitCode=0 Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.053549 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2gqx" event={"ID":"2d07947b-508d-4f12-ba1b-2d5f24a6db2c","Type":"ContainerDied","Data":"b45c0a0edb45e712987d45512f5f8181affde43a34f168d3078df992db0dd1ef"} Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.238726 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:55 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:55 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:55 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.238776 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.565673 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.611433 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kube-api-access\") pod \"5bfa0c46-abaf-4d63-964c-d8a91fafbc82\" (UID: \"5bfa0c46-abaf-4d63-964c-d8a91fafbc82\") " Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.611539 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kubelet-dir\") pod \"5bfa0c46-abaf-4d63-964c-d8a91fafbc82\" (UID: \"5bfa0c46-abaf-4d63-964c-d8a91fafbc82\") " Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.611799 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5bfa0c46-abaf-4d63-964c-d8a91fafbc82" (UID: "5bfa0c46-abaf-4d63-964c-d8a91fafbc82"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.631963 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5bfa0c46-abaf-4d63-964c-d8a91fafbc82" (UID: "5bfa0c46-abaf-4d63-964c-d8a91fafbc82"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.713245 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 11:42:55 crc kubenswrapper[4874]: I0122 11:42:55.713609 4874 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5bfa0c46-abaf-4d63-964c-d8a91fafbc82-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 11:42:56 crc kubenswrapper[4874]: I0122 11:42:56.087427 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 11:42:56 crc kubenswrapper[4874]: I0122 11:42:56.094517 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5bfa0c46-abaf-4d63-964c-d8a91fafbc82","Type":"ContainerDied","Data":"38b9f2be77edd43a5d803726e5294fbc3482629ac4f147f216e1e74861fc0fee"} Jan 22 11:42:56 crc kubenswrapper[4874]: I0122 11:42:56.094569 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38b9f2be77edd43a5d803726e5294fbc3482629ac4f147f216e1e74861fc0fee" Jan 22 11:42:56 crc kubenswrapper[4874]: I0122 11:42:56.236129 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:56 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:56 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:56 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:56 crc kubenswrapper[4874]: I0122 11:42:56.236187 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.106939 4874 generic.go:334] "Generic (PLEG): container finished" podID="0a15050a-7cbd-40f6-a656-a68293c0878a" containerID="059406f38b7eefe033761eb54f07a9cd0f802e07f88ff68ccbdb582fae9d8d28" exitCode=0 Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.106992 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" event={"ID":"0a15050a-7cbd-40f6-a656-a68293c0878a","Type":"ContainerDied","Data":"059406f38b7eefe033761eb54f07a9cd0f802e07f88ff68ccbdb582fae9d8d28"} Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.234680 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:57 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:57 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:57 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.234741 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.894533 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 22 11:42:57 crc kubenswrapper[4874]: E0122 11:42:57.894744 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bfa0c46-abaf-4d63-964c-d8a91fafbc82" containerName="pruner" Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.894756 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bfa0c46-abaf-4d63-964c-d8a91fafbc82" containerName="pruner" Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.894867 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bfa0c46-abaf-4d63-964c-d8a91fafbc82" containerName="pruner" Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.895214 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.899206 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.899330 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.913325 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.971795 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"65cb8670-9db0-44c0-9c3e-b2d669f62ba5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 11:42:57 crc kubenswrapper[4874]: I0122 11:42:57.972158 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"65cb8670-9db0-44c0-9c3e-b2d669f62ba5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.073734 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"65cb8670-9db0-44c0-9c3e-b2d669f62ba5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.073801 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"65cb8670-9db0-44c0-9c3e-b2d669f62ba5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.073876 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"65cb8670-9db0-44c0-9c3e-b2d669f62ba5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.097923 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"65cb8670-9db0-44c0-9c3e-b2d669f62ba5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.225708 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.235733 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:58 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:58 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:58 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.235785 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.595145 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.684065 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjztm\" (UniqueName: \"kubernetes.io/projected/0a15050a-7cbd-40f6-a656-a68293c0878a-kube-api-access-fjztm\") pod \"0a15050a-7cbd-40f6-a656-a68293c0878a\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.684135 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a15050a-7cbd-40f6-a656-a68293c0878a-config-volume\") pod \"0a15050a-7cbd-40f6-a656-a68293c0878a\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.684187 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a15050a-7cbd-40f6-a656-a68293c0878a-secret-volume\") pod \"0a15050a-7cbd-40f6-a656-a68293c0878a\" (UID: \"0a15050a-7cbd-40f6-a656-a68293c0878a\") " Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.687264 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a15050a-7cbd-40f6-a656-a68293c0878a-config-volume" (OuterVolumeSpecName: "config-volume") pod "0a15050a-7cbd-40f6-a656-a68293c0878a" (UID: "0a15050a-7cbd-40f6-a656-a68293c0878a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.695695 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a15050a-7cbd-40f6-a656-a68293c0878a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0a15050a-7cbd-40f6-a656-a68293c0878a" (UID: "0a15050a-7cbd-40f6-a656-a68293c0878a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.695807 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a15050a-7cbd-40f6-a656-a68293c0878a-kube-api-access-fjztm" (OuterVolumeSpecName: "kube-api-access-fjztm") pod "0a15050a-7cbd-40f6-a656-a68293c0878a" (UID: "0a15050a-7cbd-40f6-a656-a68293c0878a"). InnerVolumeSpecName "kube-api-access-fjztm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.777910 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.792793 4874 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a15050a-7cbd-40f6-a656-a68293c0878a-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.792829 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjztm\" (UniqueName: \"kubernetes.io/projected/0a15050a-7cbd-40f6-a656-a68293c0878a-kube-api-access-fjztm\") on node \"crc\" DevicePath \"\"" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.792838 4874 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a15050a-7cbd-40f6-a656-a68293c0878a-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 11:42:58 crc kubenswrapper[4874]: I0122 11:42:58.819948 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-kxvzk" Jan 22 11:42:59 crc kubenswrapper[4874]: I0122 11:42:59.168577 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" Jan 22 11:42:59 crc kubenswrapper[4874]: I0122 11:42:59.168584 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx" event={"ID":"0a15050a-7cbd-40f6-a656-a68293c0878a","Type":"ContainerDied","Data":"c99ae11d334a79b6695ceefd2cce7368234bf383f78e3abddc229aa3df93fe78"} Jan 22 11:42:59 crc kubenswrapper[4874]: I0122 11:42:59.168985 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c99ae11d334a79b6695ceefd2cce7368234bf383f78e3abddc229aa3df93fe78" Jan 22 11:42:59 crc kubenswrapper[4874]: I0122 11:42:59.171594 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"65cb8670-9db0-44c0-9c3e-b2d669f62ba5","Type":"ContainerStarted","Data":"5bec84c9490bc682a6fa2fdebd7c428a191648fb4264b6efafc24df59962bdff"} Jan 22 11:42:59 crc kubenswrapper[4874]: I0122 11:42:59.239467 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:42:59 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:42:59 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:42:59 crc kubenswrapper[4874]: healthz check failed Jan 22 11:42:59 crc kubenswrapper[4874]: I0122 11:42:59.239514 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:43:00 crc kubenswrapper[4874]: I0122 11:43:00.234741 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:43:00 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:43:00 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:43:00 crc kubenswrapper[4874]: healthz check failed Jan 22 11:43:00 crc kubenswrapper[4874]: I0122 11:43:00.234820 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:43:01 crc kubenswrapper[4874]: I0122 11:43:01.183504 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"65cb8670-9db0-44c0-9c3e-b2d669f62ba5","Type":"ContainerStarted","Data":"65f182bef3594b8ab528ba84f8dc70d785e8c2cabbc772465cad9deb5da116cd"} Jan 22 11:43:01 crc kubenswrapper[4874]: I0122 11:43:01.235686 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:43:01 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:43:01 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:43:01 crc kubenswrapper[4874]: healthz check failed Jan 22 11:43:01 crc kubenswrapper[4874]: I0122 11:43:01.235751 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:43:02 crc kubenswrapper[4874]: I0122 11:43:02.191946 4874 generic.go:334] "Generic (PLEG): container finished" podID="65cb8670-9db0-44c0-9c3e-b2d669f62ba5" containerID="65f182bef3594b8ab528ba84f8dc70d785e8c2cabbc772465cad9deb5da116cd" exitCode=0 Jan 22 11:43:02 crc kubenswrapper[4874]: I0122 11:43:02.192010 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"65cb8670-9db0-44c0-9c3e-b2d669f62ba5","Type":"ContainerDied","Data":"65f182bef3594b8ab528ba84f8dc70d785e8c2cabbc772465cad9deb5da116cd"} Jan 22 11:43:02 crc kubenswrapper[4874]: I0122 11:43:02.235411 4874 patch_prober.go:28] interesting pod/router-default-5444994796-5cczs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 11:43:02 crc kubenswrapper[4874]: [-]has-synced failed: reason withheld Jan 22 11:43:02 crc kubenswrapper[4874]: [+]process-running ok Jan 22 11:43:02 crc kubenswrapper[4874]: healthz check failed Jan 22 11:43:02 crc kubenswrapper[4874]: I0122 11:43:02.235555 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5cczs" podUID="e9446a39-9776-4de8-9137-b8952d336419" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 11:43:03 crc kubenswrapper[4874]: I0122 11:43:03.111646 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-jgw4c" Jan 22 11:43:03 crc kubenswrapper[4874]: I0122 11:43:03.189062 4874 patch_prober.go:28] interesting pod/console-f9d7485db-wws2s container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 22 11:43:03 crc kubenswrapper[4874]: I0122 11:43:03.189142 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wws2s" podUID="2dad3db6-cddd-457d-8efa-908257ef7cc5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 22 11:43:03 crc kubenswrapper[4874]: I0122 11:43:03.236115 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:43:03 crc kubenswrapper[4874]: I0122 11:43:03.240890 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-5cczs" Jan 22 11:43:06 crc kubenswrapper[4874]: I0122 11:43:06.513779 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:43:06 crc kubenswrapper[4874]: I0122 11:43:06.523662 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5451fbab-ebad-42e7-bb80-f94bad10d571-metrics-certs\") pod \"network-metrics-daemon-lr2vd\" (UID: \"5451fbab-ebad-42e7-bb80-f94bad10d571\") " pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:43:06 crc kubenswrapper[4874]: I0122 11:43:06.652200 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lr2vd" Jan 22 11:43:07 crc kubenswrapper[4874]: I0122 11:43:07.473390 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 11:43:07 crc kubenswrapper[4874]: I0122 11:43:07.629981 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kubelet-dir\") pod \"65cb8670-9db0-44c0-9c3e-b2d669f62ba5\" (UID: \"65cb8670-9db0-44c0-9c3e-b2d669f62ba5\") " Jan 22 11:43:07 crc kubenswrapper[4874]: I0122 11:43:07.630316 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kube-api-access\") pod \"65cb8670-9db0-44c0-9c3e-b2d669f62ba5\" (UID: \"65cb8670-9db0-44c0-9c3e-b2d669f62ba5\") " Jan 22 11:43:07 crc kubenswrapper[4874]: I0122 11:43:07.630152 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "65cb8670-9db0-44c0-9c3e-b2d669f62ba5" (UID: "65cb8670-9db0-44c0-9c3e-b2d669f62ba5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:43:07 crc kubenswrapper[4874]: I0122 11:43:07.633810 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "65cb8670-9db0-44c0-9c3e-b2d669f62ba5" (UID: "65cb8670-9db0-44c0-9c3e-b2d669f62ba5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:43:07 crc kubenswrapper[4874]: I0122 11:43:07.731389 4874 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 11:43:07 crc kubenswrapper[4874]: I0122 11:43:07.731451 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/65cb8670-9db0-44c0-9c3e-b2d669f62ba5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 11:43:08 crc kubenswrapper[4874]: I0122 11:43:08.238153 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"65cb8670-9db0-44c0-9c3e-b2d669f62ba5","Type":"ContainerDied","Data":"5bec84c9490bc682a6fa2fdebd7c428a191648fb4264b6efafc24df59962bdff"} Jan 22 11:43:08 crc kubenswrapper[4874]: I0122 11:43:08.238220 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bec84c9490bc682a6fa2fdebd7c428a191648fb4264b6efafc24df59962bdff" Jan 22 11:43:08 crc kubenswrapper[4874]: I0122 11:43:08.238243 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 11:43:10 crc kubenswrapper[4874]: I0122 11:43:10.242709 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:43:13 crc kubenswrapper[4874]: I0122 11:43:13.195433 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:43:13 crc kubenswrapper[4874]: I0122 11:43:13.202369 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-wws2s" Jan 22 11:43:13 crc kubenswrapper[4874]: I0122 11:43:13.521466 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:43:13 crc kubenswrapper[4874]: I0122 11:43:13.521536 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:43:15 crc kubenswrapper[4874]: I0122 11:43:15.843009 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lr2vd"] Jan 22 11:43:23 crc kubenswrapper[4874]: I0122 11:43:23.445121 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqrw2" Jan 22 11:43:25 crc kubenswrapper[4874]: E0122 11:43:25.386271 4874 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 22 11:43:25 crc kubenswrapper[4874]: E0122 11:43:25.386513 4874 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jrxzq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-pbbps_openshift-marketplace(194d73ac-fe2b-4e80-b03b-c1b780b55990): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 11:43:25 crc kubenswrapper[4874]: E0122 11:43:25.387637 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-pbbps" podUID="194d73ac-fe2b-4e80-b03b-c1b780b55990" Jan 22 11:43:26 crc kubenswrapper[4874]: E0122 11:43:26.983077 4874 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 22 11:43:26 crc kubenswrapper[4874]: E0122 11:43:26.983550 4874 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k9tqj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-bk5qd_openshift-marketplace(829e346d-eb89-4705-83c4-99d02fca8971): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 11:43:26 crc kubenswrapper[4874]: E0122 11:43:26.984725 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-bk5qd" podUID="829e346d-eb89-4705-83c4-99d02fca8971" Jan 22 11:43:30 crc kubenswrapper[4874]: E0122 11:43:30.429367 4874 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 22 11:43:30 crc kubenswrapper[4874]: E0122 11:43:30.429883 4874 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-92txx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-s2gqx_openshift-marketplace(2d07947b-508d-4f12-ba1b-2d5f24a6db2c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 11:43:30 crc kubenswrapper[4874]: E0122 11:43:30.431098 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-s2gqx" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" Jan 22 11:43:30 crc kubenswrapper[4874]: E0122 11:43:30.794457 4874 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 22 11:43:30 crc kubenswrapper[4874]: E0122 11:43:30.794730 4874 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-79nw7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xhb8k_openshift-marketplace(09a47d07-9bf6-4033-8c08-cc3aef9fe4f4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 11:43:30 crc kubenswrapper[4874]: E0122 11:43:30.795911 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xhb8k" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.094958 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 22 11:43:32 crc kubenswrapper[4874]: E0122 11:43:32.095179 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65cb8670-9db0-44c0-9c3e-b2d669f62ba5" containerName="pruner" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.095189 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="65cb8670-9db0-44c0-9c3e-b2d669f62ba5" containerName="pruner" Jan 22 11:43:32 crc kubenswrapper[4874]: E0122 11:43:32.095209 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a15050a-7cbd-40f6-a656-a68293c0878a" containerName="collect-profiles" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.095214 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a15050a-7cbd-40f6-a656-a68293c0878a" containerName="collect-profiles" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.095305 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="65cb8670-9db0-44c0-9c3e-b2d669f62ba5" containerName="pruner" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.095319 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a15050a-7cbd-40f6-a656-a68293c0878a" containerName="collect-profiles" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.095756 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.098023 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.098141 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.104295 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.267934 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57d32e28-7549-499c-a0bb-5ab789653d5e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"57d32e28-7549-499c-a0bb-5ab789653d5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.268015 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57d32e28-7549-499c-a0bb-5ab789653d5e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"57d32e28-7549-499c-a0bb-5ab789653d5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.368649 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57d32e28-7549-499c-a0bb-5ab789653d5e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"57d32e28-7549-499c-a0bb-5ab789653d5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.368723 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57d32e28-7549-499c-a0bb-5ab789653d5e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"57d32e28-7549-499c-a0bb-5ab789653d5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.368824 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57d32e28-7549-499c-a0bb-5ab789653d5e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"57d32e28-7549-499c-a0bb-5ab789653d5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.392438 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57d32e28-7549-499c-a0bb-5ab789653d5e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"57d32e28-7549-499c-a0bb-5ab789653d5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 11:43:32 crc kubenswrapper[4874]: I0122 11:43:32.413655 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 11:43:33 crc kubenswrapper[4874]: I0122 11:43:33.158219 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 11:43:34 crc kubenswrapper[4874]: E0122 11:43:34.705952 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-pbbps" podUID="194d73ac-fe2b-4e80-b03b-c1b780b55990" Jan 22 11:43:34 crc kubenswrapper[4874]: E0122 11:43:34.706278 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-bk5qd" podUID="829e346d-eb89-4705-83c4-99d02fca8971" Jan 22 11:43:34 crc kubenswrapper[4874]: E0122 11:43:34.706305 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-s2gqx" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" Jan 22 11:43:34 crc kubenswrapper[4874]: E0122 11:43:34.706348 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xhb8k" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" Jan 22 11:43:35 crc kubenswrapper[4874]: I0122 11:43:35.165312 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 22 11:43:35 crc kubenswrapper[4874]: W0122 11:43:35.177115 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod57d32e28_7549_499c_a0bb_5ab789653d5e.slice/crio-75b613d51d19f18b727cdaed48bc87cc2bc83f0df15b5099a878a65b74a0dd05 WatchSource:0}: Error finding container 75b613d51d19f18b727cdaed48bc87cc2bc83f0df15b5099a878a65b74a0dd05: Status 404 returned error can't find the container with id 75b613d51d19f18b727cdaed48bc87cc2bc83f0df15b5099a878a65b74a0dd05 Jan 22 11:43:35 crc kubenswrapper[4874]: I0122 11:43:35.389849 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"57d32e28-7549-499c-a0bb-5ab789653d5e","Type":"ContainerStarted","Data":"75b613d51d19f18b727cdaed48bc87cc2bc83f0df15b5099a878a65b74a0dd05"} Jan 22 11:43:35 crc kubenswrapper[4874]: I0122 11:43:35.391210 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" event={"ID":"5451fbab-ebad-42e7-bb80-f94bad10d571","Type":"ContainerStarted","Data":"40b8458e9937ed678c65a078d747297aea6e373c424f3a601789df0bff8b1532"} Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.613773 4874 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.614341 4874 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-djwcc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-k7sj5_openshift-marketplace(7d6b2fd5-d040-4be5-afb1-a375bddd3e88): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.615620 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-k7sj5" podUID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.670475 4874 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.670885 4874 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v87lj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-chgtv_openshift-marketplace(25f55928-a244-44f3-83a5-b6bdf551bda6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.672097 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-chgtv" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.735277 4874 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.735858 4874 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wzmfp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-cxrx9_openshift-marketplace(ede22568-7d51-42d6-b96d-1783c4e5b370): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.737597 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-cxrx9" podUID="ede22568-7d51-42d6-b96d-1783c4e5b370" Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.802567 4874 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.802728 4874 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w2xvq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-r2mwt_openshift-marketplace(dea4a6eb-c0b1-432a-81f2-e417250b0138): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 11:43:35 crc kubenswrapper[4874]: E0122 11:43:35.803971 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-r2mwt" podUID="dea4a6eb-c0b1-432a-81f2-e417250b0138" Jan 22 11:43:36 crc kubenswrapper[4874]: I0122 11:43:36.402788 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" event={"ID":"5451fbab-ebad-42e7-bb80-f94bad10d571","Type":"ContainerStarted","Data":"2a9c9d16db89a1737e501b6a2b8030392d73684bf821b2800db34795c8e53a93"} Jan 22 11:43:36 crc kubenswrapper[4874]: I0122 11:43:36.402861 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lr2vd" event={"ID":"5451fbab-ebad-42e7-bb80-f94bad10d571","Type":"ContainerStarted","Data":"bb0f7cb0546d70aa3da2091415a077df2f3e60d50033607434bef714bfd4c7a1"} Jan 22 11:43:36 crc kubenswrapper[4874]: I0122 11:43:36.407917 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"57d32e28-7549-499c-a0bb-5ab789653d5e","Type":"ContainerStarted","Data":"b0ec8f67977f1acc68b563b2d55757cc5fbf02cd92b79e7cffde786c8ba8fef4"} Jan 22 11:43:36 crc kubenswrapper[4874]: E0122 11:43:36.410507 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-cxrx9" podUID="ede22568-7d51-42d6-b96d-1783c4e5b370" Jan 22 11:43:36 crc kubenswrapper[4874]: E0122 11:43:36.410873 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-chgtv" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" Jan 22 11:43:36 crc kubenswrapper[4874]: E0122 11:43:36.411968 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-k7sj5" podUID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" Jan 22 11:43:36 crc kubenswrapper[4874]: E0122 11:43:36.412968 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-r2mwt" podUID="dea4a6eb-c0b1-432a-81f2-e417250b0138" Jan 22 11:43:36 crc kubenswrapper[4874]: I0122 11:43:36.420732 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-lr2vd" podStartSLOduration=172.420713658 podStartE2EDuration="2m52.420713658s" podCreationTimestamp="2026-01-22 11:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:43:36.415830565 +0000 UTC m=+190.260901655" watchObservedRunningTime="2026-01-22 11:43:36.420713658 +0000 UTC m=+190.265784728" Jan 22 11:43:36 crc kubenswrapper[4874]: I0122 11:43:36.892960 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 22 11:43:36 crc kubenswrapper[4874]: I0122 11:43:36.894074 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:43:36 crc kubenswrapper[4874]: I0122 11:43:36.916721 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.073869 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.073923 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b1c5453e-ced5-4d10-b696-df2a76b6a783-kube-api-access\") pod \"installer-9-crc\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.073953 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-var-lock\") pod \"installer-9-crc\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.175474 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.175537 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b1c5453e-ced5-4d10-b696-df2a76b6a783-kube-api-access\") pod \"installer-9-crc\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.175562 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-var-lock\") pod \"installer-9-crc\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.175594 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.175645 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-var-lock\") pod \"installer-9-crc\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.197574 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b1c5453e-ced5-4d10-b696-df2a76b6a783-kube-api-access\") pod \"installer-9-crc\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.216599 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.413472 4874 generic.go:334] "Generic (PLEG): container finished" podID="57d32e28-7549-499c-a0bb-5ab789653d5e" containerID="b0ec8f67977f1acc68b563b2d55757cc5fbf02cd92b79e7cffde786c8ba8fef4" exitCode=0 Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.413596 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"57d32e28-7549-499c-a0bb-5ab789653d5e","Type":"ContainerDied","Data":"b0ec8f67977f1acc68b563b2d55757cc5fbf02cd92b79e7cffde786c8ba8fef4"} Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.622490 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.687341 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.881911 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57d32e28-7549-499c-a0bb-5ab789653d5e-kube-api-access\") pod \"57d32e28-7549-499c-a0bb-5ab789653d5e\" (UID: \"57d32e28-7549-499c-a0bb-5ab789653d5e\") " Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.881984 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57d32e28-7549-499c-a0bb-5ab789653d5e-kubelet-dir\") pod \"57d32e28-7549-499c-a0bb-5ab789653d5e\" (UID: \"57d32e28-7549-499c-a0bb-5ab789653d5e\") " Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.882283 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57d32e28-7549-499c-a0bb-5ab789653d5e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "57d32e28-7549-499c-a0bb-5ab789653d5e" (UID: "57d32e28-7549-499c-a0bb-5ab789653d5e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.892150 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57d32e28-7549-499c-a0bb-5ab789653d5e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "57d32e28-7549-499c-a0bb-5ab789653d5e" (UID: "57d32e28-7549-499c-a0bb-5ab789653d5e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.983218 4874 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57d32e28-7549-499c-a0bb-5ab789653d5e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 11:43:37 crc kubenswrapper[4874]: I0122 11:43:37.983259 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57d32e28-7549-499c-a0bb-5ab789653d5e-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 11:43:38 crc kubenswrapper[4874]: I0122 11:43:38.419880 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"57d32e28-7549-499c-a0bb-5ab789653d5e","Type":"ContainerDied","Data":"75b613d51d19f18b727cdaed48bc87cc2bc83f0df15b5099a878a65b74a0dd05"} Jan 22 11:43:38 crc kubenswrapper[4874]: I0122 11:43:38.420216 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75b613d51d19f18b727cdaed48bc87cc2bc83f0df15b5099a878a65b74a0dd05" Jan 22 11:43:38 crc kubenswrapper[4874]: I0122 11:43:38.419896 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 11:43:38 crc kubenswrapper[4874]: I0122 11:43:38.420946 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b1c5453e-ced5-4d10-b696-df2a76b6a783","Type":"ContainerStarted","Data":"42a935729ff60b97150afb87bbec2f6ec94b10ae9d811cc58f762b075b789f67"} Jan 22 11:43:38 crc kubenswrapper[4874]: I0122 11:43:38.420971 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b1c5453e-ced5-4d10-b696-df2a76b6a783","Type":"ContainerStarted","Data":"ba05ca2e002e9c3237d941c91912a43cc5d0958363ae275896c78814a9a08db4"} Jan 22 11:43:38 crc kubenswrapper[4874]: I0122 11:43:38.436463 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.436376612 podStartE2EDuration="2.436376612s" podCreationTimestamp="2026-01-22 11:43:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:43:38.432884462 +0000 UTC m=+192.277955532" watchObservedRunningTime="2026-01-22 11:43:38.436376612 +0000 UTC m=+192.281447682" Jan 22 11:43:38 crc kubenswrapper[4874]: E0122 11:43:38.501202 4874 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-pod57d32e28_7549_499c_a0bb_5ab789653d5e.slice\": RecentStats: unable to find data in memory cache]" Jan 22 11:43:43 crc kubenswrapper[4874]: I0122 11:43:43.073789 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8j9tz"] Jan 22 11:43:43 crc kubenswrapper[4874]: I0122 11:43:43.520771 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:43:43 crc kubenswrapper[4874]: I0122 11:43:43.521212 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:43:48 crc kubenswrapper[4874]: I0122 11:43:48.473837 4874 generic.go:334] "Generic (PLEG): container finished" podID="194d73ac-fe2b-4e80-b03b-c1b780b55990" containerID="74aa9e987c447e71fa4ef0ece296f3c4949e2a2a7f45a155b2c483ac48eabbab" exitCode=0 Jan 22 11:43:48 crc kubenswrapper[4874]: I0122 11:43:48.473915 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbbps" event={"ID":"194d73ac-fe2b-4e80-b03b-c1b780b55990","Type":"ContainerDied","Data":"74aa9e987c447e71fa4ef0ece296f3c4949e2a2a7f45a155b2c483ac48eabbab"} Jan 22 11:43:48 crc kubenswrapper[4874]: I0122 11:43:48.477151 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bk5qd" event={"ID":"829e346d-eb89-4705-83c4-99d02fca8971","Type":"ContainerStarted","Data":"231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b"} Jan 22 11:43:48 crc kubenswrapper[4874]: I0122 11:43:48.480447 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2gqx" event={"ID":"2d07947b-508d-4f12-ba1b-2d5f24a6db2c","Type":"ContainerStarted","Data":"b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec"} Jan 22 11:43:48 crc kubenswrapper[4874]: E0122 11:43:48.629953 4874 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod829e346d_eb89_4705_83c4_99d02fca8971.slice/crio-231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod829e346d_eb89_4705_83c4_99d02fca8971.slice/crio-conmon-231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b.scope\": RecentStats: unable to find data in memory cache]" Jan 22 11:43:49 crc kubenswrapper[4874]: I0122 11:43:49.486898 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbbps" event={"ID":"194d73ac-fe2b-4e80-b03b-c1b780b55990","Type":"ContainerStarted","Data":"0648ed6c2fadca30717bac213e382fb0e0ccb45470ec390feeb8786d37e89637"} Jan 22 11:43:49 crc kubenswrapper[4874]: I0122 11:43:49.489545 4874 generic.go:334] "Generic (PLEG): container finished" podID="829e346d-eb89-4705-83c4-99d02fca8971" containerID="231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b" exitCode=0 Jan 22 11:43:49 crc kubenswrapper[4874]: I0122 11:43:49.489596 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bk5qd" event={"ID":"829e346d-eb89-4705-83c4-99d02fca8971","Type":"ContainerDied","Data":"231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b"} Jan 22 11:43:49 crc kubenswrapper[4874]: I0122 11:43:49.492973 4874 generic.go:334] "Generic (PLEG): container finished" podID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerID="b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec" exitCode=0 Jan 22 11:43:49 crc kubenswrapper[4874]: I0122 11:43:49.493044 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2gqx" event={"ID":"2d07947b-508d-4f12-ba1b-2d5f24a6db2c","Type":"ContainerDied","Data":"b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec"} Jan 22 11:43:49 crc kubenswrapper[4874]: I0122 11:43:49.501023 4874 generic.go:334] "Generic (PLEG): container finished" podID="dea4a6eb-c0b1-432a-81f2-e417250b0138" containerID="3afe6d67a395d98d3557fc317e71f0c8a6bb5a855419905a6dab04529d3efc0a" exitCode=0 Jan 22 11:43:49 crc kubenswrapper[4874]: I0122 11:43:49.501113 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2mwt" event={"ID":"dea4a6eb-c0b1-432a-81f2-e417250b0138","Type":"ContainerDied","Data":"3afe6d67a395d98d3557fc317e71f0c8a6bb5a855419905a6dab04529d3efc0a"} Jan 22 11:43:49 crc kubenswrapper[4874]: I0122 11:43:49.514343 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pbbps" podStartSLOduration=2.544358945 podStartE2EDuration="57.514323802s" podCreationTimestamp="2026-01-22 11:42:52 +0000 UTC" firstStartedPulling="2026-01-22 11:42:53.909970033 +0000 UTC m=+147.755041103" lastFinishedPulling="2026-01-22 11:43:48.87993487 +0000 UTC m=+202.725005960" observedRunningTime="2026-01-22 11:43:49.506176826 +0000 UTC m=+203.351247916" watchObservedRunningTime="2026-01-22 11:43:49.514323802 +0000 UTC m=+203.359394872" Jan 22 11:43:49 crc kubenswrapper[4874]: I0122 11:43:49.514982 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhb8k" event={"ID":"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4","Type":"ContainerStarted","Data":"0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f"} Jan 22 11:43:50 crc kubenswrapper[4874]: I0122 11:43:50.522330 4874 generic.go:334] "Generic (PLEG): container finished" podID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerID="0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f" exitCode=0 Jan 22 11:43:50 crc kubenswrapper[4874]: I0122 11:43:50.522428 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhb8k" event={"ID":"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4","Type":"ContainerDied","Data":"0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f"} Jan 22 11:43:50 crc kubenswrapper[4874]: I0122 11:43:50.524851 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxrx9" event={"ID":"ede22568-7d51-42d6-b96d-1783c4e5b370","Type":"ContainerStarted","Data":"6061b65fbd29bdf70fd1e090a65e25c9ca80adc4b7d5e07809dd238cc90776ba"} Jan 22 11:43:50 crc kubenswrapper[4874]: I0122 11:43:50.527334 4874 generic.go:334] "Generic (PLEG): container finished" podID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerID="070c1bb6dbcaa01d78ef737b95b41d8ef036a0a7e51764ac67798ef0b37e2ccf" exitCode=0 Jan 22 11:43:50 crc kubenswrapper[4874]: I0122 11:43:50.527421 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chgtv" event={"ID":"25f55928-a244-44f3-83a5-b6bdf551bda6","Type":"ContainerDied","Data":"070c1bb6dbcaa01d78ef737b95b41d8ef036a0a7e51764ac67798ef0b37e2ccf"} Jan 22 11:43:50 crc kubenswrapper[4874]: I0122 11:43:50.529194 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bk5qd" event={"ID":"829e346d-eb89-4705-83c4-99d02fca8971","Type":"ContainerStarted","Data":"051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e"} Jan 22 11:43:50 crc kubenswrapper[4874]: I0122 11:43:50.532354 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2mwt" event={"ID":"dea4a6eb-c0b1-432a-81f2-e417250b0138","Type":"ContainerStarted","Data":"1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32"} Jan 22 11:43:50 crc kubenswrapper[4874]: I0122 11:43:50.534297 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2gqx" event={"ID":"2d07947b-508d-4f12-ba1b-2d5f24a6db2c","Type":"ContainerStarted","Data":"d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd"} Jan 22 11:43:50 crc kubenswrapper[4874]: I0122 11:43:50.588293 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r2mwt" podStartSLOduration=3.334810375 podStartE2EDuration="1m1.58827092s" podCreationTimestamp="2026-01-22 11:42:49 +0000 UTC" firstStartedPulling="2026-01-22 11:42:51.742468322 +0000 UTC m=+145.587539392" lastFinishedPulling="2026-01-22 11:43:49.995928867 +0000 UTC m=+203.840999937" observedRunningTime="2026-01-22 11:43:50.584732088 +0000 UTC m=+204.429803168" watchObservedRunningTime="2026-01-22 11:43:50.58827092 +0000 UTC m=+204.433341990" Jan 22 11:43:50 crc kubenswrapper[4874]: I0122 11:43:50.605113 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bk5qd" podStartSLOduration=3.444029675 podStartE2EDuration="1m1.605093018s" podCreationTimestamp="2026-01-22 11:42:49 +0000 UTC" firstStartedPulling="2026-01-22 11:42:51.740296024 +0000 UTC m=+145.585367094" lastFinishedPulling="2026-01-22 11:43:49.901359367 +0000 UTC m=+203.746430437" observedRunningTime="2026-01-22 11:43:50.600619067 +0000 UTC m=+204.445690137" watchObservedRunningTime="2026-01-22 11:43:50.605093018 +0000 UTC m=+204.450164088" Jan 22 11:43:50 crc kubenswrapper[4874]: I0122 11:43:50.625652 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s2gqx" podStartSLOduration=3.7975601770000003 podStartE2EDuration="58.625637663s" podCreationTimestamp="2026-01-22 11:42:52 +0000 UTC" firstStartedPulling="2026-01-22 11:42:55.065568309 +0000 UTC m=+148.910639379" lastFinishedPulling="2026-01-22 11:43:49.893645785 +0000 UTC m=+203.738716865" observedRunningTime="2026-01-22 11:43:50.62493091 +0000 UTC m=+204.470001990" watchObservedRunningTime="2026-01-22 11:43:50.625637663 +0000 UTC m=+204.470708733" Jan 22 11:43:51 crc kubenswrapper[4874]: I0122 11:43:51.547580 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhb8k" event={"ID":"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4","Type":"ContainerStarted","Data":"c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a"} Jan 22 11:43:51 crc kubenswrapper[4874]: I0122 11:43:51.550115 4874 generic.go:334] "Generic (PLEG): container finished" podID="ede22568-7d51-42d6-b96d-1783c4e5b370" containerID="6061b65fbd29bdf70fd1e090a65e25c9ca80adc4b7d5e07809dd238cc90776ba" exitCode=0 Jan 22 11:43:51 crc kubenswrapper[4874]: I0122 11:43:51.550189 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxrx9" event={"ID":"ede22568-7d51-42d6-b96d-1783c4e5b370","Type":"ContainerDied","Data":"6061b65fbd29bdf70fd1e090a65e25c9ca80adc4b7d5e07809dd238cc90776ba"} Jan 22 11:43:51 crc kubenswrapper[4874]: I0122 11:43:51.553764 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chgtv" event={"ID":"25f55928-a244-44f3-83a5-b6bdf551bda6","Type":"ContainerStarted","Data":"4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7"} Jan 22 11:43:51 crc kubenswrapper[4874]: I0122 11:43:51.555250 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7sj5" event={"ID":"7d6b2fd5-d040-4be5-afb1-a375bddd3e88","Type":"ContainerStarted","Data":"55598fd3ad7b384be613bb53ff1f155b2db0dc67c238c62da8732c037ab04867"} Jan 22 11:43:51 crc kubenswrapper[4874]: I0122 11:43:51.569780 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xhb8k" podStartSLOduration=2.600058477 podStartE2EDuration="1m0.569766373s" podCreationTimestamp="2026-01-22 11:42:51 +0000 UTC" firstStartedPulling="2026-01-22 11:42:52.994598712 +0000 UTC m=+146.839669782" lastFinishedPulling="2026-01-22 11:43:50.964306618 +0000 UTC m=+204.809377678" observedRunningTime="2026-01-22 11:43:51.569485974 +0000 UTC m=+205.414557074" watchObservedRunningTime="2026-01-22 11:43:51.569766373 +0000 UTC m=+205.414837443" Jan 22 11:43:51 crc kubenswrapper[4874]: I0122 11:43:51.655698 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-chgtv" podStartSLOduration=2.792459793 podStartE2EDuration="58.655676801s" podCreationTimestamp="2026-01-22 11:42:53 +0000 UTC" firstStartedPulling="2026-01-22 11:42:55.041221321 +0000 UTC m=+148.886292391" lastFinishedPulling="2026-01-22 11:43:50.904438329 +0000 UTC m=+204.749509399" observedRunningTime="2026-01-22 11:43:51.615786449 +0000 UTC m=+205.460857529" watchObservedRunningTime="2026-01-22 11:43:51.655676801 +0000 UTC m=+205.500747871" Jan 22 11:43:52 crc kubenswrapper[4874]: I0122 11:43:52.123481 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:43:52 crc kubenswrapper[4874]: I0122 11:43:52.123576 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:43:52 crc kubenswrapper[4874]: I0122 11:43:52.539933 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:43:52 crc kubenswrapper[4874]: I0122 11:43:52.539973 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:43:52 crc kubenswrapper[4874]: I0122 11:43:52.562511 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxrx9" event={"ID":"ede22568-7d51-42d6-b96d-1783c4e5b370","Type":"ContainerStarted","Data":"88517753dde54e6943c6fe4f548a126863bb3d622e70a38576e28d6b20f09b8b"} Jan 22 11:43:52 crc kubenswrapper[4874]: I0122 11:43:52.564271 4874 generic.go:334] "Generic (PLEG): container finished" podID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" containerID="55598fd3ad7b384be613bb53ff1f155b2db0dc67c238c62da8732c037ab04867" exitCode=0 Jan 22 11:43:52 crc kubenswrapper[4874]: I0122 11:43:52.564360 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7sj5" event={"ID":"7d6b2fd5-d040-4be5-afb1-a375bddd3e88","Type":"ContainerDied","Data":"55598fd3ad7b384be613bb53ff1f155b2db0dc67c238c62da8732c037ab04867"} Jan 22 11:43:52 crc kubenswrapper[4874]: I0122 11:43:52.585148 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cxrx9" podStartSLOduration=2.36684956 podStartE2EDuration="1m2.58512973s" podCreationTimestamp="2026-01-22 11:42:50 +0000 UTC" firstStartedPulling="2026-01-22 11:42:51.738099595 +0000 UTC m=+145.583170665" lastFinishedPulling="2026-01-22 11:43:51.956379765 +0000 UTC m=+205.801450835" observedRunningTime="2026-01-22 11:43:52.583615253 +0000 UTC m=+206.428686313" watchObservedRunningTime="2026-01-22 11:43:52.58512973 +0000 UTC m=+206.430200800" Jan 22 11:43:52 crc kubenswrapper[4874]: I0122 11:43:52.608787 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:43:53 crc kubenswrapper[4874]: I0122 11:43:53.317081 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-xhb8k" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerName="registry-server" probeResult="failure" output=< Jan 22 11:43:53 crc kubenswrapper[4874]: timeout: failed to connect service ":50051" within 1s Jan 22 11:43:53 crc kubenswrapper[4874]: > Jan 22 11:43:53 crc kubenswrapper[4874]: I0122 11:43:53.348169 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:43:53 crc kubenswrapper[4874]: I0122 11:43:53.348211 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:43:53 crc kubenswrapper[4874]: I0122 11:43:53.577270 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7sj5" event={"ID":"7d6b2fd5-d040-4be5-afb1-a375bddd3e88","Type":"ContainerStarted","Data":"1cd3bbf2af5a2b7f7b7a4ab5adaef76db5460bdb370548bcbe2419923c557fc1"} Jan 22 11:43:53 crc kubenswrapper[4874]: I0122 11:43:53.601667 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k7sj5" podStartSLOduration=2.406929888 podStartE2EDuration="1m3.601651634s" podCreationTimestamp="2026-01-22 11:42:50 +0000 UTC" firstStartedPulling="2026-01-22 11:42:51.736017998 +0000 UTC m=+145.581089068" lastFinishedPulling="2026-01-22 11:43:52.930739744 +0000 UTC m=+206.775810814" observedRunningTime="2026-01-22 11:43:53.599068483 +0000 UTC m=+207.444139573" watchObservedRunningTime="2026-01-22 11:43:53.601651634 +0000 UTC m=+207.446722704" Jan 22 11:43:53 crc kubenswrapper[4874]: I0122 11:43:53.827369 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:43:53 crc kubenswrapper[4874]: I0122 11:43:53.827431 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:43:54 crc kubenswrapper[4874]: I0122 11:43:54.382185 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-s2gqx" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerName="registry-server" probeResult="failure" output=< Jan 22 11:43:54 crc kubenswrapper[4874]: timeout: failed to connect service ":50051" within 1s Jan 22 11:43:54 crc kubenswrapper[4874]: > Jan 22 11:43:54 crc kubenswrapper[4874]: I0122 11:43:54.881423 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-chgtv" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerName="registry-server" probeResult="failure" output=< Jan 22 11:43:54 crc kubenswrapper[4874]: timeout: failed to connect service ":50051" within 1s Jan 22 11:43:54 crc kubenswrapper[4874]: > Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.155620 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.155951 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.214638 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.318817 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.318922 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.388381 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.532199 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.532531 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.681471 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.746057 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.746137 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:44:00 crc kubenswrapper[4874]: I0122 11:44:00.789870 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:44:01 crc kubenswrapper[4874]: I0122 11:44:01.145122 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:44:01 crc kubenswrapper[4874]: I0122 11:44:01.145287 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:44:01 crc kubenswrapper[4874]: I0122 11:44:01.668004 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:44:01 crc kubenswrapper[4874]: I0122 11:44:01.678178 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:44:02 crc kubenswrapper[4874]: I0122 11:44:02.171888 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:44:02 crc kubenswrapper[4874]: I0122 11:44:02.242778 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:44:02 crc kubenswrapper[4874]: I0122 11:44:02.602538 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:44:02 crc kubenswrapper[4874]: I0122 11:44:02.949384 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k7sj5"] Jan 22 11:44:03 crc kubenswrapper[4874]: I0122 11:44:03.147822 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cxrx9"] Jan 22 11:44:03 crc kubenswrapper[4874]: I0122 11:44:03.416687 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:44:03 crc kubenswrapper[4874]: I0122 11:44:03.459982 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:44:03 crc kubenswrapper[4874]: I0122 11:44:03.642625 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cxrx9" podUID="ede22568-7d51-42d6-b96d-1783c4e5b370" containerName="registry-server" containerID="cri-o://88517753dde54e6943c6fe4f548a126863bb3d622e70a38576e28d6b20f09b8b" gracePeriod=2 Jan 22 11:44:03 crc kubenswrapper[4874]: I0122 11:44:03.642695 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k7sj5" podUID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" containerName="registry-server" containerID="cri-o://1cd3bbf2af5a2b7f7b7a4ab5adaef76db5460bdb370548bcbe2419923c557fc1" gracePeriod=2 Jan 22 11:44:03 crc kubenswrapper[4874]: I0122 11:44:03.890700 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:44:03 crc kubenswrapper[4874]: I0122 11:44:03.951587 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:44:05 crc kubenswrapper[4874]: I0122 11:44:05.352918 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbbps"] Jan 22 11:44:05 crc kubenswrapper[4874]: I0122 11:44:05.353574 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pbbps" podUID="194d73ac-fe2b-4e80-b03b-c1b780b55990" containerName="registry-server" containerID="cri-o://0648ed6c2fadca30717bac213e382fb0e0ccb45470ec390feeb8786d37e89637" gracePeriod=2 Jan 22 11:44:05 crc kubenswrapper[4874]: I0122 11:44:05.655076 4874 generic.go:334] "Generic (PLEG): container finished" podID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" containerID="1cd3bbf2af5a2b7f7b7a4ab5adaef76db5460bdb370548bcbe2419923c557fc1" exitCode=0 Jan 22 11:44:05 crc kubenswrapper[4874]: I0122 11:44:05.655157 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7sj5" event={"ID":"7d6b2fd5-d040-4be5-afb1-a375bddd3e88","Type":"ContainerDied","Data":"1cd3bbf2af5a2b7f7b7a4ab5adaef76db5460bdb370548bcbe2419923c557fc1"} Jan 22 11:44:06 crc kubenswrapper[4874]: I0122 11:44:06.665318 4874 generic.go:334] "Generic (PLEG): container finished" podID="ede22568-7d51-42d6-b96d-1783c4e5b370" containerID="88517753dde54e6943c6fe4f548a126863bb3d622e70a38576e28d6b20f09b8b" exitCode=0 Jan 22 11:44:06 crc kubenswrapper[4874]: I0122 11:44:06.665650 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxrx9" event={"ID":"ede22568-7d51-42d6-b96d-1783c4e5b370","Type":"ContainerDied","Data":"88517753dde54e6943c6fe4f548a126863bb3d622e70a38576e28d6b20f09b8b"} Jan 22 11:44:06 crc kubenswrapper[4874]: I0122 11:44:06.667570 4874 generic.go:334] "Generic (PLEG): container finished" podID="194d73ac-fe2b-4e80-b03b-c1b780b55990" containerID="0648ed6c2fadca30717bac213e382fb0e0ccb45470ec390feeb8786d37e89637" exitCode=0 Jan 22 11:44:06 crc kubenswrapper[4874]: I0122 11:44:06.667591 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbbps" event={"ID":"194d73ac-fe2b-4e80-b03b-c1b780b55990","Type":"ContainerDied","Data":"0648ed6c2fadca30717bac213e382fb0e0ccb45470ec390feeb8786d37e89637"} Jan 22 11:44:06 crc kubenswrapper[4874]: I0122 11:44:06.877099 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:44:06 crc kubenswrapper[4874]: I0122 11:44:06.881506 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:06.999916 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-catalog-content\") pod \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:06.999977 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-utilities\") pod \"ede22568-7d51-42d6-b96d-1783c4e5b370\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:06.999996 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-catalog-content\") pod \"ede22568-7d51-42d6-b96d-1783c4e5b370\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.000050 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djwcc\" (UniqueName: \"kubernetes.io/projected/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-kube-api-access-djwcc\") pod \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.000127 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-utilities\") pod \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\" (UID: \"7d6b2fd5-d040-4be5-afb1-a375bddd3e88\") " Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.000152 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzmfp\" (UniqueName: \"kubernetes.io/projected/ede22568-7d51-42d6-b96d-1783c4e5b370-kube-api-access-wzmfp\") pod \"ede22568-7d51-42d6-b96d-1783c4e5b370\" (UID: \"ede22568-7d51-42d6-b96d-1783c4e5b370\") " Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.000877 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-utilities" (OuterVolumeSpecName: "utilities") pod "ede22568-7d51-42d6-b96d-1783c4e5b370" (UID: "ede22568-7d51-42d6-b96d-1783c4e5b370"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.001181 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-utilities" (OuterVolumeSpecName: "utilities") pod "7d6b2fd5-d040-4be5-afb1-a375bddd3e88" (UID: "7d6b2fd5-d040-4be5-afb1-a375bddd3e88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.012790 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-kube-api-access-djwcc" (OuterVolumeSpecName: "kube-api-access-djwcc") pod "7d6b2fd5-d040-4be5-afb1-a375bddd3e88" (UID: "7d6b2fd5-d040-4be5-afb1-a375bddd3e88"). InnerVolumeSpecName "kube-api-access-djwcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.012958 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ede22568-7d51-42d6-b96d-1783c4e5b370-kube-api-access-wzmfp" (OuterVolumeSpecName: "kube-api-access-wzmfp") pod "ede22568-7d51-42d6-b96d-1783c4e5b370" (UID: "ede22568-7d51-42d6-b96d-1783c4e5b370"). InnerVolumeSpecName "kube-api-access-wzmfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.045951 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ede22568-7d51-42d6-b96d-1783c4e5b370" (UID: "ede22568-7d51-42d6-b96d-1783c4e5b370"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.051183 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d6b2fd5-d040-4be5-afb1-a375bddd3e88" (UID: "7d6b2fd5-d040-4be5-afb1-a375bddd3e88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.071980 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.101814 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.101842 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ede22568-7d51-42d6-b96d-1783c4e5b370-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.101853 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djwcc\" (UniqueName: \"kubernetes.io/projected/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-kube-api-access-djwcc\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.101861 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.101872 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzmfp\" (UniqueName: \"kubernetes.io/projected/ede22568-7d51-42d6-b96d-1783c4e5b370-kube-api-access-wzmfp\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.101880 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6b2fd5-d040-4be5-afb1-a375bddd3e88-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.202328 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-catalog-content\") pod \"194d73ac-fe2b-4e80-b03b-c1b780b55990\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.202386 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-utilities\") pod \"194d73ac-fe2b-4e80-b03b-c1b780b55990\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.202490 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrxzq\" (UniqueName: \"kubernetes.io/projected/194d73ac-fe2b-4e80-b03b-c1b780b55990-kube-api-access-jrxzq\") pod \"194d73ac-fe2b-4e80-b03b-c1b780b55990\" (UID: \"194d73ac-fe2b-4e80-b03b-c1b780b55990\") " Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.204750 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-utilities" (OuterVolumeSpecName: "utilities") pod "194d73ac-fe2b-4e80-b03b-c1b780b55990" (UID: "194d73ac-fe2b-4e80-b03b-c1b780b55990"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.206328 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/194d73ac-fe2b-4e80-b03b-c1b780b55990-kube-api-access-jrxzq" (OuterVolumeSpecName: "kube-api-access-jrxzq") pod "194d73ac-fe2b-4e80-b03b-c1b780b55990" (UID: "194d73ac-fe2b-4e80-b03b-c1b780b55990"). InnerVolumeSpecName "kube-api-access-jrxzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.227654 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "194d73ac-fe2b-4e80-b03b-c1b780b55990" (UID: "194d73ac-fe2b-4e80-b03b-c1b780b55990"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.304312 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.304354 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrxzq\" (UniqueName: \"kubernetes.io/projected/194d73ac-fe2b-4e80-b03b-c1b780b55990-kube-api-access-jrxzq\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.304421 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/194d73ac-fe2b-4e80-b03b-c1b780b55990-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.679450 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cxrx9" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.679505 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxrx9" event={"ID":"ede22568-7d51-42d6-b96d-1783c4e5b370","Type":"ContainerDied","Data":"eb9fe6572677e18214bcf1bd4f031854d94ab0f4aacf889a0900f36d0c7491fc"} Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.679603 4874 scope.go:117] "RemoveContainer" containerID="88517753dde54e6943c6fe4f548a126863bb3d622e70a38576e28d6b20f09b8b" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.685089 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbbps" event={"ID":"194d73ac-fe2b-4e80-b03b-c1b780b55990","Type":"ContainerDied","Data":"3b5a39abee6782e33316f7969db2ab23c5d711ceba05225f55e2ae3fb886847c"} Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.685235 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbbps" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.688910 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7sj5" event={"ID":"7d6b2fd5-d040-4be5-afb1-a375bddd3e88","Type":"ContainerDied","Data":"d963c7b2a3b3c57b49606220d41e2548892102c0a65c18048da0d0bded52f8de"} Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.688934 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7sj5" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.721814 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cxrx9"] Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.723681 4874 scope.go:117] "RemoveContainer" containerID="6061b65fbd29bdf70fd1e090a65e25c9ca80adc4b7d5e07809dd238cc90776ba" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.736032 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cxrx9"] Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.745036 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k7sj5"] Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.751876 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k7sj5"] Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.756344 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbbps"] Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.759713 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbbps"] Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.764929 4874 scope.go:117] "RemoveContainer" containerID="769e221159e673ec03f3822d5d4059809f4101a42a4ddda5d439bdab06ed4064" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.778376 4874 scope.go:117] "RemoveContainer" containerID="0648ed6c2fadca30717bac213e382fb0e0ccb45470ec390feeb8786d37e89637" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.792578 4874 scope.go:117] "RemoveContainer" containerID="74aa9e987c447e71fa4ef0ece296f3c4949e2a2a7f45a155b2c483ac48eabbab" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.806848 4874 scope.go:117] "RemoveContainer" containerID="1ecd914b99d6525c762a368e3ff1d7acae50f9b9d1238be939d00bb461aca834" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.822690 4874 scope.go:117] "RemoveContainer" containerID="1cd3bbf2af5a2b7f7b7a4ab5adaef76db5460bdb370548bcbe2419923c557fc1" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.867545 4874 scope.go:117] "RemoveContainer" containerID="55598fd3ad7b384be613bb53ff1f155b2db0dc67c238c62da8732c037ab04867" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.882701 4874 scope.go:117] "RemoveContainer" containerID="ab54d9fab26a3ef6698d354ee6813928ad7aedd283867428b79b6242016d11de" Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.980049 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-chgtv"] Jan 22 11:44:07 crc kubenswrapper[4874]: I0122 11:44:07.980327 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-chgtv" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerName="registry-server" containerID="cri-o://4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7" gracePeriod=2 Jan 22 11:44:08 crc kubenswrapper[4874]: I0122 11:44:08.106655 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" podUID="9fd4241f-b523-4d66-bcdb-c3bb691765c9" containerName="oauth-openshift" containerID="cri-o://0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8" gracePeriod=15 Jan 22 11:44:08 crc kubenswrapper[4874]: I0122 11:44:08.727438 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="194d73ac-fe2b-4e80-b03b-c1b780b55990" path="/var/lib/kubelet/pods/194d73ac-fe2b-4e80-b03b-c1b780b55990/volumes" Jan 22 11:44:08 crc kubenswrapper[4874]: I0122 11:44:08.728319 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" path="/var/lib/kubelet/pods/7d6b2fd5-d040-4be5-afb1-a375bddd3e88/volumes" Jan 22 11:44:08 crc kubenswrapper[4874]: I0122 11:44:08.728907 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ede22568-7d51-42d6-b96d-1783c4e5b370" path="/var/lib/kubelet/pods/ede22568-7d51-42d6-b96d-1783c4e5b370/volumes" Jan 22 11:44:08 crc kubenswrapper[4874]: E0122 11:44:08.872785 4874 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25f55928_a244_44f3_83a5_b6bdf551bda6.slice/crio-conmon-4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7.scope\": RecentStats: unable to find data in memory cache]" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.478247 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.586172 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.633837 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-catalog-content\") pod \"25f55928-a244-44f3-83a5-b6bdf551bda6\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.633876 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-utilities\") pod \"25f55928-a244-44f3-83a5-b6bdf551bda6\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.633972 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v87lj\" (UniqueName: \"kubernetes.io/projected/25f55928-a244-44f3-83a5-b6bdf551bda6-kube-api-access-v87lj\") pod \"25f55928-a244-44f3-83a5-b6bdf551bda6\" (UID: \"25f55928-a244-44f3-83a5-b6bdf551bda6\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.635876 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-utilities" (OuterVolumeSpecName: "utilities") pod "25f55928-a244-44f3-83a5-b6bdf551bda6" (UID: "25f55928-a244-44f3-83a5-b6bdf551bda6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.641532 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25f55928-a244-44f3-83a5-b6bdf551bda6-kube-api-access-v87lj" (OuterVolumeSpecName: "kube-api-access-v87lj") pod "25f55928-a244-44f3-83a5-b6bdf551bda6" (UID: "25f55928-a244-44f3-83a5-b6bdf551bda6"). InnerVolumeSpecName "kube-api-access-v87lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.659620 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-df7774cfb-gpckj"] Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.659865 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d32e28-7549-499c-a0bb-5ab789653d5e" containerName="pruner" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.659878 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d32e28-7549-499c-a0bb-5ab789653d5e" containerName="pruner" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.659909 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ede22568-7d51-42d6-b96d-1783c4e5b370" containerName="extract-content" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.659916 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="ede22568-7d51-42d6-b96d-1783c4e5b370" containerName="extract-content" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.659926 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="194d73ac-fe2b-4e80-b03b-c1b780b55990" containerName="extract-content" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.659932 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="194d73ac-fe2b-4e80-b03b-c1b780b55990" containerName="extract-content" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.659939 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd4241f-b523-4d66-bcdb-c3bb691765c9" containerName="oauth-openshift" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.659945 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd4241f-b523-4d66-bcdb-c3bb691765c9" containerName="oauth-openshift" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.659952 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="194d73ac-fe2b-4e80-b03b-c1b780b55990" containerName="extract-utilities" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.659959 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="194d73ac-fe2b-4e80-b03b-c1b780b55990" containerName="extract-utilities" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.659990 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" containerName="extract-content" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.659996 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" containerName="extract-content" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.660003 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660009 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.660017 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="194d73ac-fe2b-4e80-b03b-c1b780b55990" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660022 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="194d73ac-fe2b-4e80-b03b-c1b780b55990" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.660030 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerName="extract-content" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660036 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerName="extract-content" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.660062 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerName="extract-utilities" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660068 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerName="extract-utilities" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.660077 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ede22568-7d51-42d6-b96d-1783c4e5b370" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660083 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="ede22568-7d51-42d6-b96d-1783c4e5b370" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.660089 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ede22568-7d51-42d6-b96d-1783c4e5b370" containerName="extract-utilities" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660095 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="ede22568-7d51-42d6-b96d-1783c4e5b370" containerName="extract-utilities" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.660103 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" containerName="extract-utilities" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660109 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" containerName="extract-utilities" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.660118 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660142 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660246 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d6b2fd5-d040-4be5-afb1-a375bddd3e88" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660256 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="57d32e28-7549-499c-a0bb-5ab789653d5e" containerName="pruner" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660265 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="194d73ac-fe2b-4e80-b03b-c1b780b55990" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660272 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fd4241f-b523-4d66-bcdb-c3bb691765c9" containerName="oauth-openshift" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660300 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660309 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="ede22568-7d51-42d6-b96d-1783c4e5b370" containerName="registry-server" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.660680 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.677782 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-df7774cfb-gpckj"] Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.718348 4874 generic.go:334] "Generic (PLEG): container finished" podID="9fd4241f-b523-4d66-bcdb-c3bb691765c9" containerID="0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8" exitCode=0 Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.718453 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" event={"ID":"9fd4241f-b523-4d66-bcdb-c3bb691765c9","Type":"ContainerDied","Data":"0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8"} Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.718675 4874 scope.go:117] "RemoveContainer" containerID="0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.718773 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" event={"ID":"9fd4241f-b523-4d66-bcdb-c3bb691765c9","Type":"ContainerDied","Data":"ebdd6071b5dc3f0f50c15dd7a89f7157d85c2d83e8eb328a3a8f9e41e0ceff64"} Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.719164 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8j9tz" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.723761 4874 generic.go:334] "Generic (PLEG): container finished" podID="25f55928-a244-44f3-83a5-b6bdf551bda6" containerID="4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7" exitCode=0 Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.723804 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chgtv" event={"ID":"25f55928-a244-44f3-83a5-b6bdf551bda6","Type":"ContainerDied","Data":"4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7"} Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.723854 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chgtv" event={"ID":"25f55928-a244-44f3-83a5-b6bdf551bda6","Type":"ContainerDied","Data":"f695035bc56b57dcd6c7f8fc2e3f5c5198706ed0269a97d26b4dab2f6db84183"} Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.723860 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-chgtv" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.735736 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-router-certs\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.735809 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-serving-cert\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736136 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-ocp-branding-template\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736176 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-session\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736222 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-service-ca\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736251 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-cliconfig\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736281 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-policies\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736330 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-idp-0-file-data\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736362 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-dir\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736423 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-provider-selection\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736460 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-error\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736498 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9gf8\" (UniqueName: \"kubernetes.io/projected/9fd4241f-b523-4d66-bcdb-c3bb691765c9-kube-api-access-h9gf8\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736530 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-trusted-ca-bundle\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736566 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-login\") pod \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\" (UID: \"9fd4241f-b523-4d66-bcdb-c3bb691765c9\") " Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736748 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-audit-policies\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736808 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dd041a63-dbb7-4dd5-b42f-f715f4686487-audit-dir\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736831 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-template-login\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736868 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-service-ca\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736890 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-cliconfig\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.736968 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.737004 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-serving-cert\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.737051 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-template-error\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.737072 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.737102 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.737128 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-router-certs\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.737152 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-session\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.737179 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6m9d\" (UniqueName: \"kubernetes.io/projected/dd041a63-dbb7-4dd5-b42f-f715f4686487-kube-api-access-c6m9d\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.737200 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.737248 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.737264 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v87lj\" (UniqueName: \"kubernetes.io/projected/25f55928-a244-44f3-83a5-b6bdf551bda6-kube-api-access-v87lj\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.737815 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.738459 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.738584 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.739057 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.739121 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.739598 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.742558 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.742737 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.742917 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.743164 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.743833 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.744093 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.744197 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.744346 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fd4241f-b523-4d66-bcdb-c3bb691765c9-kube-api-access-h9gf8" (OuterVolumeSpecName: "kube-api-access-h9gf8") pod "9fd4241f-b523-4d66-bcdb-c3bb691765c9" (UID: "9fd4241f-b523-4d66-bcdb-c3bb691765c9"). InnerVolumeSpecName "kube-api-access-h9gf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.748678 4874 scope.go:117] "RemoveContainer" containerID="0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.749161 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8\": container with ID starting with 0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8 not found: ID does not exist" containerID="0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.749200 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8"} err="failed to get container status \"0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8\": rpc error: code = NotFound desc = could not find container \"0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8\": container with ID starting with 0846f2eb4e8eb047949a104105735cd3c58d1f0d30fc25a2c6f738a80fe669e8 not found: ID does not exist" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.749253 4874 scope.go:117] "RemoveContainer" containerID="4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.762363 4874 scope.go:117] "RemoveContainer" containerID="070c1bb6dbcaa01d78ef737b95b41d8ef036a0a7e51764ac67798ef0b37e2ccf" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.771213 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25f55928-a244-44f3-83a5-b6bdf551bda6" (UID: "25f55928-a244-44f3-83a5-b6bdf551bda6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.777624 4874 scope.go:117] "RemoveContainer" containerID="a94610599ca8c298d3b9c45ba5eaf696e08b7bd8dfa7298e41a50594597af12d" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.797297 4874 scope.go:117] "RemoveContainer" containerID="4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.797763 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7\": container with ID starting with 4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7 not found: ID does not exist" containerID="4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.797806 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7"} err="failed to get container status \"4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7\": rpc error: code = NotFound desc = could not find container \"4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7\": container with ID starting with 4e6bb093ab81592869bb3d215fdba956487ddaa72974f7be113130776c97fcd7 not found: ID does not exist" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.797840 4874 scope.go:117] "RemoveContainer" containerID="070c1bb6dbcaa01d78ef737b95b41d8ef036a0a7e51764ac67798ef0b37e2ccf" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.798137 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"070c1bb6dbcaa01d78ef737b95b41d8ef036a0a7e51764ac67798ef0b37e2ccf\": container with ID starting with 070c1bb6dbcaa01d78ef737b95b41d8ef036a0a7e51764ac67798ef0b37e2ccf not found: ID does not exist" containerID="070c1bb6dbcaa01d78ef737b95b41d8ef036a0a7e51764ac67798ef0b37e2ccf" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.798182 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070c1bb6dbcaa01d78ef737b95b41d8ef036a0a7e51764ac67798ef0b37e2ccf"} err="failed to get container status \"070c1bb6dbcaa01d78ef737b95b41d8ef036a0a7e51764ac67798ef0b37e2ccf\": rpc error: code = NotFound desc = could not find container \"070c1bb6dbcaa01d78ef737b95b41d8ef036a0a7e51764ac67798ef0b37e2ccf\": container with ID starting with 070c1bb6dbcaa01d78ef737b95b41d8ef036a0a7e51764ac67798ef0b37e2ccf not found: ID does not exist" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.798212 4874 scope.go:117] "RemoveContainer" containerID="a94610599ca8c298d3b9c45ba5eaf696e08b7bd8dfa7298e41a50594597af12d" Jan 22 11:44:09 crc kubenswrapper[4874]: E0122 11:44:09.798664 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a94610599ca8c298d3b9c45ba5eaf696e08b7bd8dfa7298e41a50594597af12d\": container with ID starting with a94610599ca8c298d3b9c45ba5eaf696e08b7bd8dfa7298e41a50594597af12d not found: ID does not exist" containerID="a94610599ca8c298d3b9c45ba5eaf696e08b7bd8dfa7298e41a50594597af12d" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.798723 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a94610599ca8c298d3b9c45ba5eaf696e08b7bd8dfa7298e41a50594597af12d"} err="failed to get container status \"a94610599ca8c298d3b9c45ba5eaf696e08b7bd8dfa7298e41a50594597af12d\": rpc error: code = NotFound desc = could not find container \"a94610599ca8c298d3b9c45ba5eaf696e08b7bd8dfa7298e41a50594597af12d\": container with ID starting with a94610599ca8c298d3b9c45ba5eaf696e08b7bd8dfa7298e41a50594597af12d not found: ID does not exist" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838179 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838236 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-serving-cert\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838268 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-template-error\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838293 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838327 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838358 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-router-certs\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838390 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-session\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838449 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6m9d\" (UniqueName: \"kubernetes.io/projected/dd041a63-dbb7-4dd5-b42f-f715f4686487-kube-api-access-c6m9d\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838483 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838529 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-audit-policies\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838619 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dd041a63-dbb7-4dd5-b42f-f715f4686487-audit-dir\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838650 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-template-login\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838699 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-service-ca\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838733 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-cliconfig\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838808 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838827 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838844 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838862 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838879 4874 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838895 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f55928-a244-44f3-83a5-b6bdf551bda6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838913 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838929 4874 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9fd4241f-b523-4d66-bcdb-c3bb691765c9-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838948 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838968 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.838985 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9gf8\" (UniqueName: \"kubernetes.io/projected/9fd4241f-b523-4d66-bcdb-c3bb691765c9-kube-api-access-h9gf8\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.839001 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.839022 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.839038 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.839053 4874 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9fd4241f-b523-4d66-bcdb-c3bb691765c9-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.839600 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dd041a63-dbb7-4dd5-b42f-f715f4686487-audit-dir\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.839893 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-cliconfig\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.840193 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.840349 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-service-ca\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.841213 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dd041a63-dbb7-4dd5-b42f-f715f4686487-audit-policies\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.842519 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-template-error\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.842638 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-router-certs\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.843702 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-template-login\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.843828 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.845555 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-serving-cert\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.845763 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.845864 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.847644 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dd041a63-dbb7-4dd5-b42f-f715f4686487-v4-0-config-system-session\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.866638 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6m9d\" (UniqueName: \"kubernetes.io/projected/dd041a63-dbb7-4dd5-b42f-f715f4686487-kube-api-access-c6m9d\") pod \"oauth-openshift-df7774cfb-gpckj\" (UID: \"dd041a63-dbb7-4dd5-b42f-f715f4686487\") " pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:09 crc kubenswrapper[4874]: I0122 11:44:09.994345 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:10 crc kubenswrapper[4874]: I0122 11:44:10.098235 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-chgtv"] Jan 22 11:44:10 crc kubenswrapper[4874]: I0122 11:44:10.104209 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-chgtv"] Jan 22 11:44:10 crc kubenswrapper[4874]: I0122 11:44:10.111220 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8j9tz"] Jan 22 11:44:10 crc kubenswrapper[4874]: I0122 11:44:10.111270 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8j9tz"] Jan 22 11:44:10 crc kubenswrapper[4874]: I0122 11:44:10.463588 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-df7774cfb-gpckj"] Jan 22 11:44:10 crc kubenswrapper[4874]: W0122 11:44:10.474476 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd041a63_dbb7_4dd5_b42f_f715f4686487.slice/crio-2e609717995262733d45fa73c905d6bf626b78e039ed91e72db579d70bf67f64 WatchSource:0}: Error finding container 2e609717995262733d45fa73c905d6bf626b78e039ed91e72db579d70bf67f64: Status 404 returned error can't find the container with id 2e609717995262733d45fa73c905d6bf626b78e039ed91e72db579d70bf67f64 Jan 22 11:44:10 crc kubenswrapper[4874]: I0122 11:44:10.729687 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25f55928-a244-44f3-83a5-b6bdf551bda6" path="/var/lib/kubelet/pods/25f55928-a244-44f3-83a5-b6bdf551bda6/volumes" Jan 22 11:44:10 crc kubenswrapper[4874]: I0122 11:44:10.730632 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fd4241f-b523-4d66-bcdb-c3bb691765c9" path="/var/lib/kubelet/pods/9fd4241f-b523-4d66-bcdb-c3bb691765c9/volumes" Jan 22 11:44:10 crc kubenswrapper[4874]: I0122 11:44:10.733760 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" event={"ID":"dd041a63-dbb7-4dd5-b42f-f715f4686487","Type":"ContainerStarted","Data":"2e609717995262733d45fa73c905d6bf626b78e039ed91e72db579d70bf67f64"} Jan 22 11:44:11 crc kubenswrapper[4874]: I0122 11:44:11.739518 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" event={"ID":"dd041a63-dbb7-4dd5-b42f-f715f4686487","Type":"ContainerStarted","Data":"00324d0b7f08ac0455ed071de4cfd597cfb0cb30bd337ca733fbb70cd91bdbd9"} Jan 22 11:44:11 crc kubenswrapper[4874]: I0122 11:44:11.739874 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:11 crc kubenswrapper[4874]: I0122 11:44:11.744844 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" Jan 22 11:44:11 crc kubenswrapper[4874]: I0122 11:44:11.759759 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-df7774cfb-gpckj" podStartSLOduration=28.759736195 podStartE2EDuration="28.759736195s" podCreationTimestamp="2026-01-22 11:43:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:44:11.75642232 +0000 UTC m=+225.601493410" watchObservedRunningTime="2026-01-22 11:44:11.759736195 +0000 UTC m=+225.604807265" Jan 22 11:44:13 crc kubenswrapper[4874]: I0122 11:44:13.520538 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:44:13 crc kubenswrapper[4874]: I0122 11:44:13.520662 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:44:13 crc kubenswrapper[4874]: I0122 11:44:13.520731 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:44:13 crc kubenswrapper[4874]: I0122 11:44:13.521644 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 11:44:13 crc kubenswrapper[4874]: I0122 11:44:13.521751 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192" gracePeriod=600 Jan 22 11:44:13 crc kubenswrapper[4874]: I0122 11:44:13.763761 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192" exitCode=0 Jan 22 11:44:13 crc kubenswrapper[4874]: I0122 11:44:13.763824 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192"} Jan 22 11:44:14 crc kubenswrapper[4874]: I0122 11:44:14.776714 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"cf675adc04e058930041f77cfb016f23f15475800ec1dca3cd6db1579e71257b"} Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.613632 4874 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.614639 4874 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.614802 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.615019 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb" gracePeriod=15 Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.615079 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d" gracePeriod=15 Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.615013 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e" gracePeriod=15 Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.615168 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88" gracePeriod=15 Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.615152 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea" gracePeriod=15 Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.615733 4874 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 11:44:15 crc kubenswrapper[4874]: E0122 11:44:15.615928 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.615949 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 11:44:15 crc kubenswrapper[4874]: E0122 11:44:15.615965 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.615974 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 22 11:44:15 crc kubenswrapper[4874]: E0122 11:44:15.615986 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.615995 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 22 11:44:15 crc kubenswrapper[4874]: E0122 11:44:15.616008 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.616016 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 22 11:44:15 crc kubenswrapper[4874]: E0122 11:44:15.616029 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.616038 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 22 11:44:15 crc kubenswrapper[4874]: E0122 11:44:15.616054 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.616064 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 11:44:15 crc kubenswrapper[4874]: E0122 11:44:15.616074 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.616083 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.616208 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.616221 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.616233 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.616244 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.616258 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.616267 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.725739 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.725787 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.725873 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.725903 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.725923 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.725954 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.725985 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.726031 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.783113 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.785023 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.785913 4874 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e" exitCode=0 Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.785945 4874 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88" exitCode=0 Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.785954 4874 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d" exitCode=0 Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.785961 4874 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea" exitCode=2 Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.786034 4874 scope.go:117] "RemoveContainer" containerID="ef9dd0379be6fda0ab37fd98b2f8118e8ee772c2293b312689a2f14816024782" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.826763 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.826835 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.826856 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.826888 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.826909 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.826960 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.826968 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.826918 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.827003 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.827058 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.827066 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.827094 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.827216 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.827323 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.827333 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:15 crc kubenswrapper[4874]: I0122 11:44:15.827463 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:16 crc kubenswrapper[4874]: I0122 11:44:16.721286 4874 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:16 crc kubenswrapper[4874]: I0122 11:44:16.796363 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 11:44:16 crc kubenswrapper[4874]: I0122 11:44:16.800291 4874 generic.go:334] "Generic (PLEG): container finished" podID="b1c5453e-ced5-4d10-b696-df2a76b6a783" containerID="42a935729ff60b97150afb87bbec2f6ec94b10ae9d811cc58f762b075b789f67" exitCode=0 Jan 22 11:44:16 crc kubenswrapper[4874]: I0122 11:44:16.800368 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b1c5453e-ced5-4d10-b696-df2a76b6a783","Type":"ContainerDied","Data":"42a935729ff60b97150afb87bbec2f6ec94b10ae9d811cc58f762b075b789f67"} Jan 22 11:44:16 crc kubenswrapper[4874]: I0122 11:44:16.801427 4874 status_manager.go:851] "Failed to get status for pod" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:17 crc kubenswrapper[4874]: E0122 11:44:17.753238 4874 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" volumeName="registry-storage" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.033154 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.034098 4874 status_manager.go:851] "Failed to get status for pod" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.057933 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b1c5453e-ced5-4d10-b696-df2a76b6a783-kube-api-access\") pod \"b1c5453e-ced5-4d10-b696-df2a76b6a783\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.058059 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-kubelet-dir\") pod \"b1c5453e-ced5-4d10-b696-df2a76b6a783\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.058101 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-var-lock\") pod \"b1c5453e-ced5-4d10-b696-df2a76b6a783\" (UID: \"b1c5453e-ced5-4d10-b696-df2a76b6a783\") " Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.058552 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b1c5453e-ced5-4d10-b696-df2a76b6a783" (UID: "b1c5453e-ced5-4d10-b696-df2a76b6a783"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.058572 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-var-lock" (OuterVolumeSpecName: "var-lock") pod "b1c5453e-ced5-4d10-b696-df2a76b6a783" (UID: "b1c5453e-ced5-4d10-b696-df2a76b6a783"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.063688 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1c5453e-ced5-4d10-b696-df2a76b6a783-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b1c5453e-ced5-4d10-b696-df2a76b6a783" (UID: "b1c5453e-ced5-4d10-b696-df2a76b6a783"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.098194 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.099190 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.100054 4874 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.101863 4874 status_manager.go:851] "Failed to get status for pod" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.159488 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.159884 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.160159 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.159695 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.159945 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.160269 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.161885 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b1c5453e-ced5-4d10-b696-df2a76b6a783-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.162076 4874 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.162210 4874 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b1c5453e-ced5-4d10-b696-df2a76b6a783-var-lock\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.162339 4874 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.162496 4874 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.162640 4874 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.726874 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.815306 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b1c5453e-ced5-4d10-b696-df2a76b6a783","Type":"ContainerDied","Data":"ba05ca2e002e9c3237d941c91912a43cc5d0958363ae275896c78814a9a08db4"} Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.815696 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba05ca2e002e9c3237d941c91912a43cc5d0958363ae275896c78814a9a08db4" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.815374 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.818802 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.819116 4874 status_manager.go:851] "Failed to get status for pod" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.820387 4874 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb" exitCode=0 Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.820479 4874 scope.go:117] "RemoveContainer" containerID="18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.820537 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.821498 4874 status_manager.go:851] "Failed to get status for pod" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.821716 4874 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.824042 4874 status_manager.go:851] "Failed to get status for pod" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.824214 4874 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.848955 4874 scope.go:117] "RemoveContainer" containerID="d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.863371 4874 scope.go:117] "RemoveContainer" containerID="9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.877829 4874 scope.go:117] "RemoveContainer" containerID="1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.890452 4874 scope.go:117] "RemoveContainer" containerID="cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.912869 4874 scope.go:117] "RemoveContainer" containerID="a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.929414 4874 scope.go:117] "RemoveContainer" containerID="18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e" Jan 22 11:44:18 crc kubenswrapper[4874]: E0122 11:44:18.929845 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\": container with ID starting with 18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e not found: ID does not exist" containerID="18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.929891 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e"} err="failed to get container status \"18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\": rpc error: code = NotFound desc = could not find container \"18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e\": container with ID starting with 18ef891e18831de897e651335dbab8089da981d5b7ed6663fd98fc78e2f9a78e not found: ID does not exist" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.929920 4874 scope.go:117] "RemoveContainer" containerID="d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88" Jan 22 11:44:18 crc kubenswrapper[4874]: E0122 11:44:18.930344 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\": container with ID starting with d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88 not found: ID does not exist" containerID="d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.930374 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88"} err="failed to get container status \"d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\": rpc error: code = NotFound desc = could not find container \"d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88\": container with ID starting with d67ea1437424158afdd37b9c6c83aa949691a12bfb9fc02403b84782f8581c88 not found: ID does not exist" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.930392 4874 scope.go:117] "RemoveContainer" containerID="9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d" Jan 22 11:44:18 crc kubenswrapper[4874]: E0122 11:44:18.934076 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\": container with ID starting with 9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d not found: ID does not exist" containerID="9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.934143 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d"} err="failed to get container status \"9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\": rpc error: code = NotFound desc = could not find container \"9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d\": container with ID starting with 9df7a5a3d92759dcbbee7531933f0acb22a64c50f61c755349b62b014b1efd1d not found: ID does not exist" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.934195 4874 scope.go:117] "RemoveContainer" containerID="1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea" Jan 22 11:44:18 crc kubenswrapper[4874]: E0122 11:44:18.934685 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\": container with ID starting with 1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea not found: ID does not exist" containerID="1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.934753 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea"} err="failed to get container status \"1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\": rpc error: code = NotFound desc = could not find container \"1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea\": container with ID starting with 1db3356e896765e6aa09a9812c36a172116aec98395c8a9901937823555795ea not found: ID does not exist" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.934786 4874 scope.go:117] "RemoveContainer" containerID="cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb" Jan 22 11:44:18 crc kubenswrapper[4874]: E0122 11:44:18.935153 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\": container with ID starting with cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb not found: ID does not exist" containerID="cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.935180 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb"} err="failed to get container status \"cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\": rpc error: code = NotFound desc = could not find container \"cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb\": container with ID starting with cae6eda3d4ce208d450cd707291f2aeafb2b580500e3bd85fadf6e079dec9ceb not found: ID does not exist" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.935196 4874 scope.go:117] "RemoveContainer" containerID="a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1" Jan 22 11:44:18 crc kubenswrapper[4874]: E0122 11:44:18.935622 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\": container with ID starting with a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1 not found: ID does not exist" containerID="a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1" Jan 22 11:44:18 crc kubenswrapper[4874]: I0122 11:44:18.935660 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1"} err="failed to get container status \"a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\": rpc error: code = NotFound desc = could not find container \"a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1\": container with ID starting with a8fc1c5897cb1005db092fb9109a9687fa2d55c7bdd3e81cc8b2cbb68f01dbe1 not found: ID does not exist" Jan 22 11:44:20 crc kubenswrapper[4874]: E0122 11:44:20.654578 4874 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:20 crc kubenswrapper[4874]: I0122 11:44:20.655273 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:20 crc kubenswrapper[4874]: E0122 11:44:20.686167 4874 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.153:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188d0af16e9c6bbb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 11:44:20.685614011 +0000 UTC m=+234.530685091,LastTimestamp:2026-01-22 11:44:20.685614011 +0000 UTC m=+234.530685091,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 22 11:44:20 crc kubenswrapper[4874]: I0122 11:44:20.833977 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"e49a6e4f262e8fee29b4eb5addf749ce3f57dfda52784b0c23b853a78fbd7d52"} Jan 22 11:44:21 crc kubenswrapper[4874]: I0122 11:44:21.841413 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7"} Jan 22 11:44:21 crc kubenswrapper[4874]: E0122 11:44:21.842200 4874 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:21 crc kubenswrapper[4874]: I0122 11:44:21.842438 4874 status_manager.go:851] "Failed to get status for pod" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:21 crc kubenswrapper[4874]: E0122 11:44:21.856754 4874 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:21 crc kubenswrapper[4874]: E0122 11:44:21.856923 4874 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:21 crc kubenswrapper[4874]: E0122 11:44:21.857173 4874 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:21 crc kubenswrapper[4874]: E0122 11:44:21.857615 4874 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:21 crc kubenswrapper[4874]: E0122 11:44:21.858097 4874 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:21 crc kubenswrapper[4874]: I0122 11:44:21.858124 4874 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 22 11:44:21 crc kubenswrapper[4874]: E0122 11:44:21.858532 4874 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="200ms" Jan 22 11:44:22 crc kubenswrapper[4874]: E0122 11:44:22.059517 4874 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="400ms" Jan 22 11:44:22 crc kubenswrapper[4874]: E0122 11:44:22.460914 4874 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="800ms" Jan 22 11:44:22 crc kubenswrapper[4874]: E0122 11:44:22.849255 4874 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:44:23 crc kubenswrapper[4874]: E0122 11:44:23.071100 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:44:23Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:44:23Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:44:23Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T11:44:23Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:23 crc kubenswrapper[4874]: E0122 11:44:23.071768 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:23 crc kubenswrapper[4874]: E0122 11:44:23.072231 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:23 crc kubenswrapper[4874]: E0122 11:44:23.072537 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:23 crc kubenswrapper[4874]: E0122 11:44:23.072981 4874 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:23 crc kubenswrapper[4874]: E0122 11:44:23.073011 4874 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 11:44:23 crc kubenswrapper[4874]: E0122 11:44:23.262373 4874 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="1.6s" Jan 22 11:44:24 crc kubenswrapper[4874]: E0122 11:44:24.863729 4874 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="3.2s" Jan 22 11:44:25 crc kubenswrapper[4874]: E0122 11:44:25.626721 4874 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.153:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188d0af16e9c6bbb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 11:44:20.685614011 +0000 UTC m=+234.530685091,LastTimestamp:2026-01-22 11:44:20.685614011 +0000 UTC m=+234.530685091,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 22 11:44:26 crc kubenswrapper[4874]: I0122 11:44:26.720051 4874 status_manager.go:851] "Failed to get status for pod" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:27 crc kubenswrapper[4874]: I0122 11:44:27.715144 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:27 crc kubenswrapper[4874]: I0122 11:44:27.716394 4874 status_manager.go:851] "Failed to get status for pod" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:27 crc kubenswrapper[4874]: I0122 11:44:27.738356 4874 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d78bedba-59c5-4a8a-89c2-4414b15e80a7" Jan 22 11:44:27 crc kubenswrapper[4874]: I0122 11:44:27.738441 4874 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d78bedba-59c5-4a8a-89c2-4414b15e80a7" Jan 22 11:44:27 crc kubenswrapper[4874]: E0122 11:44:27.738971 4874 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:27 crc kubenswrapper[4874]: I0122 11:44:27.739629 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:27 crc kubenswrapper[4874]: W0122 11:44:27.778779 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-f3488b4f2fc617bfa2217b39cd9de34b2d359400f65ba521d8ccda0f4051f4b6 WatchSource:0}: Error finding container f3488b4f2fc617bfa2217b39cd9de34b2d359400f65ba521d8ccda0f4051f4b6: Status 404 returned error can't find the container with id f3488b4f2fc617bfa2217b39cd9de34b2d359400f65ba521d8ccda0f4051f4b6 Jan 22 11:44:27 crc kubenswrapper[4874]: I0122 11:44:27.880234 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f3488b4f2fc617bfa2217b39cd9de34b2d359400f65ba521d8ccda0f4051f4b6"} Jan 22 11:44:28 crc kubenswrapper[4874]: E0122 11:44:28.065338 4874 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="6.4s" Jan 22 11:44:28 crc kubenswrapper[4874]: I0122 11:44:28.888298 4874 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="fd0a1ae174bc904842bd2821353b4de9bffa6edc8f0e47e624553679d131e045" exitCode=0 Jan 22 11:44:28 crc kubenswrapper[4874]: I0122 11:44:28.888379 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"fd0a1ae174bc904842bd2821353b4de9bffa6edc8f0e47e624553679d131e045"} Jan 22 11:44:28 crc kubenswrapper[4874]: I0122 11:44:28.888704 4874 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d78bedba-59c5-4a8a-89c2-4414b15e80a7" Jan 22 11:44:28 crc kubenswrapper[4874]: I0122 11:44:28.888766 4874 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d78bedba-59c5-4a8a-89c2-4414b15e80a7" Jan 22 11:44:28 crc kubenswrapper[4874]: E0122 11:44:28.889268 4874 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:28 crc kubenswrapper[4874]: I0122 11:44:28.889324 4874 status_manager.go:851] "Failed to get status for pod" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Jan 22 11:44:29 crc kubenswrapper[4874]: E0122 11:44:29.113340 4874 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793.scope\": RecentStats: unable to find data in memory cache]" Jan 22 11:44:29 crc kubenswrapper[4874]: I0122 11:44:29.909213 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 22 11:44:29 crc kubenswrapper[4874]: I0122 11:44:29.909573 4874 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793" exitCode=1 Jan 22 11:44:29 crc kubenswrapper[4874]: I0122 11:44:29.909628 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793"} Jan 22 11:44:29 crc kubenswrapper[4874]: I0122 11:44:29.910065 4874 scope.go:117] "RemoveContainer" containerID="6d285287ccdbe99a7eed164ef995aab026c7088d89680574ad45a7f1fe24a793" Jan 22 11:44:29 crc kubenswrapper[4874]: I0122 11:44:29.922257 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ce3b1f8042cf273d0bcd53a00c902433a2ca31502d16f33d515aa09c8365791c"} Jan 22 11:44:29 crc kubenswrapper[4874]: I0122 11:44:29.922300 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d98d6525724c83080512bcbd4c829515e848026a175beb67959da73f4a62f304"} Jan 22 11:44:29 crc kubenswrapper[4874]: I0122 11:44:29.922317 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"df3c94ef87e3ca81fb4b8203fe451b7723686c560ceb594466daef05d9d8504b"} Jan 22 11:44:29 crc kubenswrapper[4874]: I0122 11:44:29.922328 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"107dd77e930b33f518438417063b199a81c13785679029c5025e4ac7dec019fd"} Jan 22 11:44:30 crc kubenswrapper[4874]: I0122 11:44:30.929629 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 22 11:44:30 crc kubenswrapper[4874]: I0122 11:44:30.929981 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0713db3892c7dafcf5a96b65709f3c24ec50d410cf96df8562bfe588a57586c4"} Jan 22 11:44:30 crc kubenswrapper[4874]: I0122 11:44:30.932889 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5a781d7d94c394dd791a3ac354d65cfece4064d9db0f75ee05ad177d791a33ee"} Jan 22 11:44:30 crc kubenswrapper[4874]: I0122 11:44:30.933089 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:30 crc kubenswrapper[4874]: I0122 11:44:30.933188 4874 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d78bedba-59c5-4a8a-89c2-4414b15e80a7" Jan 22 11:44:30 crc kubenswrapper[4874]: I0122 11:44:30.933206 4874 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d78bedba-59c5-4a8a-89c2-4414b15e80a7" Jan 22 11:44:32 crc kubenswrapper[4874]: I0122 11:44:32.630904 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 11:44:32 crc kubenswrapper[4874]: I0122 11:44:32.739827 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:32 crc kubenswrapper[4874]: I0122 11:44:32.739921 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:32 crc kubenswrapper[4874]: I0122 11:44:32.747003 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:34 crc kubenswrapper[4874]: I0122 11:44:34.904946 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 11:44:34 crc kubenswrapper[4874]: I0122 11:44:34.906064 4874 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 22 11:44:34 crc kubenswrapper[4874]: I0122 11:44:34.906115 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 22 11:44:35 crc kubenswrapper[4874]: I0122 11:44:35.941128 4874 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:35 crc kubenswrapper[4874]: I0122 11:44:35.957712 4874 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d78bedba-59c5-4a8a-89c2-4414b15e80a7" Jan 22 11:44:35 crc kubenswrapper[4874]: I0122 11:44:35.957743 4874 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d78bedba-59c5-4a8a-89c2-4414b15e80a7" Jan 22 11:44:35 crc kubenswrapper[4874]: I0122 11:44:35.961813 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:36 crc kubenswrapper[4874]: I0122 11:44:36.736327 4874 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="9cb63f2a-a5c9-41c5-ab57-dcdfc7e42870" Jan 22 11:44:36 crc kubenswrapper[4874]: I0122 11:44:36.970099 4874 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d78bedba-59c5-4a8a-89c2-4414b15e80a7" Jan 22 11:44:36 crc kubenswrapper[4874]: I0122 11:44:36.970466 4874 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d78bedba-59c5-4a8a-89c2-4414b15e80a7" Jan 22 11:44:36 crc kubenswrapper[4874]: I0122 11:44:36.977059 4874 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="9cb63f2a-a5c9-41c5-ab57-dcdfc7e42870" Jan 22 11:44:44 crc kubenswrapper[4874]: I0122 11:44:44.904647 4874 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 22 11:44:44 crc kubenswrapper[4874]: I0122 11:44:44.905519 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 22 11:44:46 crc kubenswrapper[4874]: I0122 11:44:46.005379 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 22 11:44:46 crc kubenswrapper[4874]: I0122 11:44:46.744901 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 22 11:44:46 crc kubenswrapper[4874]: I0122 11:44:46.830443 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 22 11:44:47 crc kubenswrapper[4874]: I0122 11:44:47.143799 4874 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 22 11:44:47 crc kubenswrapper[4874]: I0122 11:44:47.146666 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 22 11:44:47 crc kubenswrapper[4874]: I0122 11:44:47.378490 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 22 11:44:47 crc kubenswrapper[4874]: I0122 11:44:47.476919 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 22 11:44:47 crc kubenswrapper[4874]: I0122 11:44:47.548045 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 22 11:44:47 crc kubenswrapper[4874]: I0122 11:44:47.904584 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 22 11:44:48 crc kubenswrapper[4874]: I0122 11:44:48.521387 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 22 11:44:48 crc kubenswrapper[4874]: I0122 11:44:48.584377 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 22 11:44:48 crc kubenswrapper[4874]: I0122 11:44:48.656854 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 22 11:44:48 crc kubenswrapper[4874]: I0122 11:44:48.737483 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 22 11:44:48 crc kubenswrapper[4874]: I0122 11:44:48.972756 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 22 11:44:48 crc kubenswrapper[4874]: I0122 11:44:48.981182 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.017030 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.021229 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.231672 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.282292 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.420169 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.439283 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.510420 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.599869 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.605144 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.718024 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.768259 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.768496 4874 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 22 11:44:49 crc kubenswrapper[4874]: I0122 11:44:49.959043 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.163970 4874 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.178861 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.211600 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.234311 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.237870 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.239040 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.241004 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.260942 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.633822 4874 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.639655 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.639726 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.654707 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.662158 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.667648 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.676909 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=15.676887675 podStartE2EDuration="15.676887675s" podCreationTimestamp="2026-01-22 11:44:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:44:50.672339352 +0000 UTC m=+264.517410462" watchObservedRunningTime="2026-01-22 11:44:50.676887675 +0000 UTC m=+264.521958755" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.805219 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.812255 4874 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.825547 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.852764 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.877712 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.890571 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.912774 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.917910 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 22 11:44:50 crc kubenswrapper[4874]: I0122 11:44:50.925762 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.053621 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.062513 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.076060 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.144871 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.239211 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.288369 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.380557 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.391472 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.395338 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.473074 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.486471 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.489111 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.544278 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.555991 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.612920 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.769532 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.790500 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.874481 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.899571 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.900897 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.993345 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 22 11:44:51 crc kubenswrapper[4874]: I0122 11:44:51.996693 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.024024 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.087881 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.130473 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.179376 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.182876 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.244808 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.267092 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.381374 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.527458 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.558571 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.561225 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.659206 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.738329 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.802704 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.827202 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.912532 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.937492 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.978341 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 22 11:44:52 crc kubenswrapper[4874]: I0122 11:44:52.985451 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.005140 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.005457 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.057488 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.077727 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.123263 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.126156 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.137368 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.156657 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.168838 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.191218 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.201860 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.227171 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.233213 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.316757 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.325780 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.478776 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.499589 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.534334 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.684673 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.702435 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.702902 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.717541 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.751423 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.760862 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.791203 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.851200 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.857007 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.942910 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.979476 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 22 11:44:53 crc kubenswrapper[4874]: I0122 11:44:53.984409 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.000008 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.087950 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.213633 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.214450 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.244766 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.317439 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.319465 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.385482 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.421682 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.448040 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.570666 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.659155 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.694479 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.845118 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.910605 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 11:44:54 crc kubenswrapper[4874]: I0122 11:44:54.916200 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.032476 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.137988 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.187683 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.228548 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.276615 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.315053 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.316314 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.348100 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.379556 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.655104 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.733042 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.886997 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 22 11:44:55 crc kubenswrapper[4874]: I0122 11:44:55.903743 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.067678 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.090691 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.112202 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.200361 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.203080 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.225729 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.250708 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.347740 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.421425 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.502288 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.515770 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.578135 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.706686 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.895056 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 22 11:44:56 crc kubenswrapper[4874]: I0122 11:44:56.945093 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.002271 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.113316 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.128991 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.137840 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.167937 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.190420 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.225738 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.252769 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.261954 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.350993 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.401165 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.402003 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.450370 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.510095 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.565741 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.571160 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.605039 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.612477 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.721682 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.783945 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.843530 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.921963 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.958832 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 22 11:44:57 crc kubenswrapper[4874]: I0122 11:44:57.969512 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.185230 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.295053 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.408230 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.417437 4874 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.427933 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.433576 4874 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.433809 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7" gracePeriod=5 Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.450431 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.477563 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.554143 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.614100 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.631734 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.705133 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.781918 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.805331 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.815127 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.848191 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.875714 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 22 11:44:58 crc kubenswrapper[4874]: I0122 11:44:58.891199 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.035429 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.221746 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.248643 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.304004 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.349628 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.351915 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.369043 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.371284 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.438117 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.477528 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.477531 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.550582 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.551881 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.601538 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.677637 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.729861 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.745824 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.776170 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.860891 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 22 11:44:59 crc kubenswrapper[4874]: I0122 11:44:59.949274 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 11:45:00 crc kubenswrapper[4874]: I0122 11:45:00.138308 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 22 11:45:00 crc kubenswrapper[4874]: I0122 11:45:00.248206 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 22 11:45:00 crc kubenswrapper[4874]: I0122 11:45:00.321500 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 22 11:45:00 crc kubenswrapper[4874]: I0122 11:45:00.494622 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 22 11:45:00 crc kubenswrapper[4874]: I0122 11:45:00.495159 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 22 11:45:00 crc kubenswrapper[4874]: I0122 11:45:00.817489 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 22 11:45:00 crc kubenswrapper[4874]: I0122 11:45:00.869170 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 22 11:45:01 crc kubenswrapper[4874]: I0122 11:45:01.091998 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 22 11:45:01 crc kubenswrapper[4874]: I0122 11:45:01.184648 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 22 11:45:01 crc kubenswrapper[4874]: I0122 11:45:01.251108 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 22 11:45:01 crc kubenswrapper[4874]: I0122 11:45:01.293889 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 22 11:45:01 crc kubenswrapper[4874]: I0122 11:45:01.329444 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 22 11:45:01 crc kubenswrapper[4874]: I0122 11:45:01.384959 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 22 11:45:01 crc kubenswrapper[4874]: I0122 11:45:01.403192 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 11:45:01 crc kubenswrapper[4874]: I0122 11:45:01.449784 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 22 11:45:01 crc kubenswrapper[4874]: I0122 11:45:01.560804 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 22 11:45:01 crc kubenswrapper[4874]: I0122 11:45:01.893893 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 22 11:45:01 crc kubenswrapper[4874]: I0122 11:45:01.899273 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 22 11:45:02 crc kubenswrapper[4874]: I0122 11:45:02.152950 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 22 11:45:02 crc kubenswrapper[4874]: I0122 11:45:02.378898 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 22 11:45:02 crc kubenswrapper[4874]: I0122 11:45:02.397924 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 22 11:45:02 crc kubenswrapper[4874]: I0122 11:45:02.449597 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 22 11:45:02 crc kubenswrapper[4874]: I0122 11:45:02.500856 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 22 11:45:02 crc kubenswrapper[4874]: I0122 11:45:02.573470 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 22 11:45:02 crc kubenswrapper[4874]: I0122 11:45:02.659568 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 22 11:45:02 crc kubenswrapper[4874]: I0122 11:45:02.704748 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 22 11:45:03 crc kubenswrapper[4874]: I0122 11:45:03.205471 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 22 11:45:03 crc kubenswrapper[4874]: I0122 11:45:03.359467 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 22 11:45:03 crc kubenswrapper[4874]: I0122 11:45:03.686292 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.000325 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.000389 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.112501 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.112580 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.112598 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.112667 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.112719 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.113052 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.113318 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.113332 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.113434 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.119571 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.123034 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.123101 4874 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7" exitCode=137 Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.123160 4874 scope.go:117] "RemoveContainer" containerID="7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.123211 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.154876 4874 scope.go:117] "RemoveContainer" containerID="7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7" Jan 22 11:45:04 crc kubenswrapper[4874]: E0122 11:45:04.155376 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7\": container with ID starting with 7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7 not found: ID does not exist" containerID="7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.155438 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7"} err="failed to get container status \"7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7\": rpc error: code = NotFound desc = could not find container \"7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7\": container with ID starting with 7b60b861ffc1abe331ac35bff6d3cb66212140990112c08598f13b185feceff7 not found: ID does not exist" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.214483 4874 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.214516 4874 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.214525 4874 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.214533 4874 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.214541 4874 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.632983 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 22 11:45:04 crc kubenswrapper[4874]: I0122 11:45:04.728149 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.714458 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn"] Jan 22 11:45:11 crc kubenswrapper[4874]: E0122 11:45:11.715581 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.715599 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 22 11:45:11 crc kubenswrapper[4874]: E0122 11:45:11.715612 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" containerName="installer" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.715619 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" containerName="installer" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.715741 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.715755 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1c5453e-ced5-4d10-b696-df2a76b6a783" containerName="installer" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.716656 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.719077 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.719181 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.731599 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn"] Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.811368 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce45c6ff-601e-4b12-97b6-4737304db2d7-config-volume\") pod \"collect-profiles-29484705-xlhwn\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.811681 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txdrs\" (UniqueName: \"kubernetes.io/projected/ce45c6ff-601e-4b12-97b6-4737304db2d7-kube-api-access-txdrs\") pod \"collect-profiles-29484705-xlhwn\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.811778 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce45c6ff-601e-4b12-97b6-4737304db2d7-secret-volume\") pod \"collect-profiles-29484705-xlhwn\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.913198 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txdrs\" (UniqueName: \"kubernetes.io/projected/ce45c6ff-601e-4b12-97b6-4737304db2d7-kube-api-access-txdrs\") pod \"collect-profiles-29484705-xlhwn\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.913587 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce45c6ff-601e-4b12-97b6-4737304db2d7-secret-volume\") pod \"collect-profiles-29484705-xlhwn\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.913670 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce45c6ff-601e-4b12-97b6-4737304db2d7-config-volume\") pod \"collect-profiles-29484705-xlhwn\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.914628 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce45c6ff-601e-4b12-97b6-4737304db2d7-config-volume\") pod \"collect-profiles-29484705-xlhwn\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.919218 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce45c6ff-601e-4b12-97b6-4737304db2d7-secret-volume\") pod \"collect-profiles-29484705-xlhwn\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:11 crc kubenswrapper[4874]: I0122 11:45:11.940291 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txdrs\" (UniqueName: \"kubernetes.io/projected/ce45c6ff-601e-4b12-97b6-4737304db2d7-kube-api-access-txdrs\") pod \"collect-profiles-29484705-xlhwn\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:12 crc kubenswrapper[4874]: I0122 11:45:12.033359 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:12 crc kubenswrapper[4874]: I0122 11:45:12.402838 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn"] Jan 22 11:45:13 crc kubenswrapper[4874]: I0122 11:45:13.179709 4874 generic.go:334] "Generic (PLEG): container finished" podID="ce45c6ff-601e-4b12-97b6-4737304db2d7" containerID="e4f9ca4f65548ceb4a0c2ab28cc0dcb70d2bfed6abf0e18366dac77436253610" exitCode=0 Jan 22 11:45:13 crc kubenswrapper[4874]: I0122 11:45:13.180023 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" event={"ID":"ce45c6ff-601e-4b12-97b6-4737304db2d7","Type":"ContainerDied","Data":"e4f9ca4f65548ceb4a0c2ab28cc0dcb70d2bfed6abf0e18366dac77436253610"} Jan 22 11:45:13 crc kubenswrapper[4874]: I0122 11:45:13.180063 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" event={"ID":"ce45c6ff-601e-4b12-97b6-4737304db2d7","Type":"ContainerStarted","Data":"0ad7d6ea5a30e76ca03ea0e0ab963e27f80e9b8ce1ebe3c88dec81c8a4abfc22"} Jan 22 11:45:14 crc kubenswrapper[4874]: I0122 11:45:14.390331 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:14 crc kubenswrapper[4874]: I0122 11:45:14.545081 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce45c6ff-601e-4b12-97b6-4737304db2d7-secret-volume\") pod \"ce45c6ff-601e-4b12-97b6-4737304db2d7\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " Jan 22 11:45:14 crc kubenswrapper[4874]: I0122 11:45:14.545457 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txdrs\" (UniqueName: \"kubernetes.io/projected/ce45c6ff-601e-4b12-97b6-4737304db2d7-kube-api-access-txdrs\") pod \"ce45c6ff-601e-4b12-97b6-4737304db2d7\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " Jan 22 11:45:14 crc kubenswrapper[4874]: I0122 11:45:14.545664 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce45c6ff-601e-4b12-97b6-4737304db2d7-config-volume\") pod \"ce45c6ff-601e-4b12-97b6-4737304db2d7\" (UID: \"ce45c6ff-601e-4b12-97b6-4737304db2d7\") " Jan 22 11:45:14 crc kubenswrapper[4874]: I0122 11:45:14.546177 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce45c6ff-601e-4b12-97b6-4737304db2d7-config-volume" (OuterVolumeSpecName: "config-volume") pod "ce45c6ff-601e-4b12-97b6-4737304db2d7" (UID: "ce45c6ff-601e-4b12-97b6-4737304db2d7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:45:14 crc kubenswrapper[4874]: I0122 11:45:14.550035 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce45c6ff-601e-4b12-97b6-4737304db2d7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ce45c6ff-601e-4b12-97b6-4737304db2d7" (UID: "ce45c6ff-601e-4b12-97b6-4737304db2d7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:45:14 crc kubenswrapper[4874]: I0122 11:45:14.550331 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce45c6ff-601e-4b12-97b6-4737304db2d7-kube-api-access-txdrs" (OuterVolumeSpecName: "kube-api-access-txdrs") pod "ce45c6ff-601e-4b12-97b6-4737304db2d7" (UID: "ce45c6ff-601e-4b12-97b6-4737304db2d7"). InnerVolumeSpecName "kube-api-access-txdrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:45:14 crc kubenswrapper[4874]: I0122 11:45:14.647037 4874 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce45c6ff-601e-4b12-97b6-4737304db2d7-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:14 crc kubenswrapper[4874]: I0122 11:45:14.647083 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txdrs\" (UniqueName: \"kubernetes.io/projected/ce45c6ff-601e-4b12-97b6-4737304db2d7-kube-api-access-txdrs\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:14 crc kubenswrapper[4874]: I0122 11:45:14.647102 4874 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce45c6ff-601e-4b12-97b6-4737304db2d7-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:15 crc kubenswrapper[4874]: I0122 11:45:15.190114 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" event={"ID":"ce45c6ff-601e-4b12-97b6-4737304db2d7","Type":"ContainerDied","Data":"0ad7d6ea5a30e76ca03ea0e0ab963e27f80e9b8ce1ebe3c88dec81c8a4abfc22"} Jan 22 11:45:15 crc kubenswrapper[4874]: I0122 11:45:15.190428 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ad7d6ea5a30e76ca03ea0e0ab963e27f80e9b8ce1ebe3c88dec81c8a4abfc22" Jan 22 11:45:15 crc kubenswrapper[4874]: I0122 11:45:15.190204 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn" Jan 22 11:45:20 crc kubenswrapper[4874]: I0122 11:45:20.225382 4874 generic.go:334] "Generic (PLEG): container finished" podID="213d34f5-75cd-459c-9e56-2938fe5e3950" containerID="a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3" exitCode=0 Jan 22 11:45:20 crc kubenswrapper[4874]: I0122 11:45:20.225644 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" event={"ID":"213d34f5-75cd-459c-9e56-2938fe5e3950","Type":"ContainerDied","Data":"a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3"} Jan 22 11:45:20 crc kubenswrapper[4874]: I0122 11:45:20.227723 4874 scope.go:117] "RemoveContainer" containerID="a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3" Jan 22 11:45:21 crc kubenswrapper[4874]: I0122 11:45:21.236599 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" event={"ID":"213d34f5-75cd-459c-9e56-2938fe5e3950","Type":"ContainerStarted","Data":"a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb"} Jan 22 11:45:21 crc kubenswrapper[4874]: I0122 11:45:21.237369 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:45:21 crc kubenswrapper[4874]: I0122 11:45:21.243495 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.399838 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-swcnq"] Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.400556 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" podUID="f5464a23-ec80-4717-bfe0-6efeab811853" containerName="controller-manager" containerID="cri-o://fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5" gracePeriod=30 Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.484108 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87"] Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.484358 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" podUID="65f71c2e-ab34-4d33-905f-609555dab78c" containerName="route-controller-manager" containerID="cri-o://84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1" gracePeriod=30 Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.839732 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.895680 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.964387 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m29d2\" (UniqueName: \"kubernetes.io/projected/65f71c2e-ab34-4d33-905f-609555dab78c-kube-api-access-m29d2\") pod \"65f71c2e-ab34-4d33-905f-609555dab78c\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.964501 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nm6v\" (UniqueName: \"kubernetes.io/projected/f5464a23-ec80-4717-bfe0-6efeab811853-kube-api-access-2nm6v\") pod \"f5464a23-ec80-4717-bfe0-6efeab811853\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.964551 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-proxy-ca-bundles\") pod \"f5464a23-ec80-4717-bfe0-6efeab811853\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.964588 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-config\") pod \"f5464a23-ec80-4717-bfe0-6efeab811853\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.964616 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65f71c2e-ab34-4d33-905f-609555dab78c-serving-cert\") pod \"65f71c2e-ab34-4d33-905f-609555dab78c\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.964652 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-client-ca\") pod \"65f71c2e-ab34-4d33-905f-609555dab78c\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.964684 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5464a23-ec80-4717-bfe0-6efeab811853-serving-cert\") pod \"f5464a23-ec80-4717-bfe0-6efeab811853\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.964829 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-client-ca\") pod \"f5464a23-ec80-4717-bfe0-6efeab811853\" (UID: \"f5464a23-ec80-4717-bfe0-6efeab811853\") " Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.965522 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f5464a23-ec80-4717-bfe0-6efeab811853" (UID: "f5464a23-ec80-4717-bfe0-6efeab811853"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.965540 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-config" (OuterVolumeSpecName: "config") pod "f5464a23-ec80-4717-bfe0-6efeab811853" (UID: "f5464a23-ec80-4717-bfe0-6efeab811853"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.965577 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-client-ca" (OuterVolumeSpecName: "client-ca") pod "f5464a23-ec80-4717-bfe0-6efeab811853" (UID: "f5464a23-ec80-4717-bfe0-6efeab811853"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.965680 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-config\") pod \"65f71c2e-ab34-4d33-905f-609555dab78c\" (UID: \"65f71c2e-ab34-4d33-905f-609555dab78c\") " Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.966233 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-client-ca" (OuterVolumeSpecName: "client-ca") pod "65f71c2e-ab34-4d33-905f-609555dab78c" (UID: "65f71c2e-ab34-4d33-905f-609555dab78c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.966853 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-config" (OuterVolumeSpecName: "config") pod "65f71c2e-ab34-4d33-905f-609555dab78c" (UID: "65f71c2e-ab34-4d33-905f-609555dab78c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.967568 4874 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.967596 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.967609 4874 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.967628 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5464a23-ec80-4717-bfe0-6efeab811853-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.967638 4874 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65f71c2e-ab34-4d33-905f-609555dab78c-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.970530 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5464a23-ec80-4717-bfe0-6efeab811853-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f5464a23-ec80-4717-bfe0-6efeab811853" (UID: "f5464a23-ec80-4717-bfe0-6efeab811853"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.970701 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65f71c2e-ab34-4d33-905f-609555dab78c-kube-api-access-m29d2" (OuterVolumeSpecName: "kube-api-access-m29d2") pod "65f71c2e-ab34-4d33-905f-609555dab78c" (UID: "65f71c2e-ab34-4d33-905f-609555dab78c"). InnerVolumeSpecName "kube-api-access-m29d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.970698 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5464a23-ec80-4717-bfe0-6efeab811853-kube-api-access-2nm6v" (OuterVolumeSpecName: "kube-api-access-2nm6v") pod "f5464a23-ec80-4717-bfe0-6efeab811853" (UID: "f5464a23-ec80-4717-bfe0-6efeab811853"). InnerVolumeSpecName "kube-api-access-2nm6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:45:23 crc kubenswrapper[4874]: I0122 11:45:23.970727 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f71c2e-ab34-4d33-905f-609555dab78c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "65f71c2e-ab34-4d33-905f-609555dab78c" (UID: "65f71c2e-ab34-4d33-905f-609555dab78c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.069246 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m29d2\" (UniqueName: \"kubernetes.io/projected/65f71c2e-ab34-4d33-905f-609555dab78c-kube-api-access-m29d2\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.069279 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nm6v\" (UniqueName: \"kubernetes.io/projected/f5464a23-ec80-4717-bfe0-6efeab811853-kube-api-access-2nm6v\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.069288 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65f71c2e-ab34-4d33-905f-609555dab78c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.069297 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5464a23-ec80-4717-bfe0-6efeab811853-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.256943 4874 generic.go:334] "Generic (PLEG): container finished" podID="65f71c2e-ab34-4d33-905f-609555dab78c" containerID="84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1" exitCode=0 Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.257025 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" event={"ID":"65f71c2e-ab34-4d33-905f-609555dab78c","Type":"ContainerDied","Data":"84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1"} Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.257057 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" event={"ID":"65f71c2e-ab34-4d33-905f-609555dab78c","Type":"ContainerDied","Data":"a9c651399a50d57c614df95905597a481669874ba9ca8205cf5d6198908fbe7c"} Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.257059 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.257076 4874 scope.go:117] "RemoveContainer" containerID="84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.260672 4874 generic.go:334] "Generic (PLEG): container finished" podID="f5464a23-ec80-4717-bfe0-6efeab811853" containerID="fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5" exitCode=0 Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.260746 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" event={"ID":"f5464a23-ec80-4717-bfe0-6efeab811853","Type":"ContainerDied","Data":"fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5"} Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.260770 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.260797 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-swcnq" event={"ID":"f5464a23-ec80-4717-bfe0-6efeab811853","Type":"ContainerDied","Data":"d437fa062da853d4db086da0d6818e76b56a3a9b163a7e48bb221915ade96982"} Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.282646 4874 scope.go:117] "RemoveContainer" containerID="84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1" Jan 22 11:45:24 crc kubenswrapper[4874]: E0122 11:45:24.283214 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1\": container with ID starting with 84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1 not found: ID does not exist" containerID="84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.283252 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1"} err="failed to get container status \"84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1\": rpc error: code = NotFound desc = could not find container \"84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1\": container with ID starting with 84a7330ee9f1d0d95ea8e0bf8a052aae4a2dfa5a3e264d704406ce13a297fbb1 not found: ID does not exist" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.283280 4874 scope.go:117] "RemoveContainer" containerID="fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.311301 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87"] Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.316975 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2bn87"] Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.327773 4874 scope.go:117] "RemoveContainer" containerID="fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5" Jan 22 11:45:24 crc kubenswrapper[4874]: E0122 11:45:24.328251 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5\": container with ID starting with fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5 not found: ID does not exist" containerID="fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.328311 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-swcnq"] Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.328303 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5"} err="failed to get container status \"fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5\": rpc error: code = NotFound desc = could not find container \"fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5\": container with ID starting with fff471b9ef5cb3368fda0285ee54baef17701d5ccdb0a7c909cc9a9b7b3f50f5 not found: ID does not exist" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.338462 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-swcnq"] Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.707195 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2"] Jan 22 11:45:24 crc kubenswrapper[4874]: E0122 11:45:24.708099 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce45c6ff-601e-4b12-97b6-4737304db2d7" containerName="collect-profiles" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.708133 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce45c6ff-601e-4b12-97b6-4737304db2d7" containerName="collect-profiles" Jan 22 11:45:24 crc kubenswrapper[4874]: E0122 11:45:24.708160 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f71c2e-ab34-4d33-905f-609555dab78c" containerName="route-controller-manager" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.708177 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f71c2e-ab34-4d33-905f-609555dab78c" containerName="route-controller-manager" Jan 22 11:45:24 crc kubenswrapper[4874]: E0122 11:45:24.708212 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5464a23-ec80-4717-bfe0-6efeab811853" containerName="controller-manager" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.708231 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5464a23-ec80-4717-bfe0-6efeab811853" containerName="controller-manager" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.708487 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5464a23-ec80-4717-bfe0-6efeab811853" containerName="controller-manager" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.708520 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce45c6ff-601e-4b12-97b6-4737304db2d7" containerName="collect-profiles" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.708540 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f71c2e-ab34-4d33-905f-609555dab78c" containerName="route-controller-manager" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.709181 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.711594 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.711636 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7"] Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.711656 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.711841 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.712337 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.713384 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.714153 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.716589 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.720033 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.720737 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.721756 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.722233 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.724125 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.724571 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.734357 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65f71c2e-ab34-4d33-905f-609555dab78c" path="/var/lib/kubelet/pods/65f71c2e-ab34-4d33-905f-609555dab78c/volumes" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.735238 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5464a23-ec80-4717-bfe0-6efeab811853" path="/var/lib/kubelet/pods/f5464a23-ec80-4717-bfe0-6efeab811853/volumes" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.735875 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7"] Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.739688 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2"] Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.783198 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-client-ca\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.783246 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-client-ca\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.783274 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e9625e0-ef53-4afd-9576-a41ad312a97d-serving-cert\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.783303 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xklr5\" (UniqueName: \"kubernetes.io/projected/8ba4583a-60a2-4ace-ba80-8d804556d604-kube-api-access-xklr5\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.783322 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-config\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.783406 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-proxy-ca-bundles\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.783750 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rmmk\" (UniqueName: \"kubernetes.io/projected/9e9625e0-ef53-4afd-9576-a41ad312a97d-kube-api-access-2rmmk\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.783848 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ba4583a-60a2-4ace-ba80-8d804556d604-serving-cert\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.783922 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-config\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.788052 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.885737 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-client-ca\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.885819 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-client-ca\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.885857 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e9625e0-ef53-4afd-9576-a41ad312a97d-serving-cert\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.885896 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xklr5\" (UniqueName: \"kubernetes.io/projected/8ba4583a-60a2-4ace-ba80-8d804556d604-kube-api-access-xklr5\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.885925 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-config\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.885947 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-proxy-ca-bundles\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.885992 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rmmk\" (UniqueName: \"kubernetes.io/projected/9e9625e0-ef53-4afd-9576-a41ad312a97d-kube-api-access-2rmmk\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.886017 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ba4583a-60a2-4ace-ba80-8d804556d604-serving-cert\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.886045 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-config\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.887027 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-client-ca\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.887986 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-config\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.888294 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-proxy-ca-bundles\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.889112 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-config\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.889899 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-client-ca\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.891257 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ba4583a-60a2-4ace-ba80-8d804556d604-serving-cert\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.892226 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e9625e0-ef53-4afd-9576-a41ad312a97d-serving-cert\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.905674 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rmmk\" (UniqueName: \"kubernetes.io/projected/9e9625e0-ef53-4afd-9576-a41ad312a97d-kube-api-access-2rmmk\") pod \"controller-manager-789bbbcf9f-jm8f7\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:24 crc kubenswrapper[4874]: I0122 11:45:24.908187 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xklr5\" (UniqueName: \"kubernetes.io/projected/8ba4583a-60a2-4ace-ba80-8d804556d604-kube-api-access-xklr5\") pod \"route-controller-manager-5764494d47-vsfq2\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:25 crc kubenswrapper[4874]: I0122 11:45:25.086763 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:25 crc kubenswrapper[4874]: I0122 11:45:25.095713 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:25 crc kubenswrapper[4874]: I0122 11:45:25.305950 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2"] Jan 22 11:45:25 crc kubenswrapper[4874]: I0122 11:45:25.332067 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7"] Jan 22 11:45:25 crc kubenswrapper[4874]: W0122 11:45:25.334995 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e9625e0_ef53_4afd_9576_a41ad312a97d.slice/crio-d2b018a92f66b0200304d7741f10b85f1e493ae49ca6a4372b8a9605031feb95 WatchSource:0}: Error finding container d2b018a92f66b0200304d7741f10b85f1e493ae49ca6a4372b8a9605031feb95: Status 404 returned error can't find the container with id d2b018a92f66b0200304d7741f10b85f1e493ae49ca6a4372b8a9605031feb95 Jan 22 11:45:26 crc kubenswrapper[4874]: I0122 11:45:26.277289 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" event={"ID":"8ba4583a-60a2-4ace-ba80-8d804556d604","Type":"ContainerStarted","Data":"21043c10d05ec41a76bd77364b386c886df31d414e1e8c2604a1659796f8c27e"} Jan 22 11:45:26 crc kubenswrapper[4874]: I0122 11:45:26.277972 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:26 crc kubenswrapper[4874]: I0122 11:45:26.277991 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" event={"ID":"8ba4583a-60a2-4ace-ba80-8d804556d604","Type":"ContainerStarted","Data":"ef570dea7a83f0ce123b85b79bfbc070103a700770f9817e620b420f82ef7f8a"} Jan 22 11:45:26 crc kubenswrapper[4874]: I0122 11:45:26.279539 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" event={"ID":"9e9625e0-ef53-4afd-9576-a41ad312a97d","Type":"ContainerStarted","Data":"5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983"} Jan 22 11:45:26 crc kubenswrapper[4874]: I0122 11:45:26.279654 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" event={"ID":"9e9625e0-ef53-4afd-9576-a41ad312a97d","Type":"ContainerStarted","Data":"d2b018a92f66b0200304d7741f10b85f1e493ae49ca6a4372b8a9605031feb95"} Jan 22 11:45:26 crc kubenswrapper[4874]: I0122 11:45:26.279883 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:26 crc kubenswrapper[4874]: I0122 11:45:26.283482 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:26 crc kubenswrapper[4874]: I0122 11:45:26.292345 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:45:26 crc kubenswrapper[4874]: I0122 11:45:26.296751 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" podStartSLOduration=3.296735142 podStartE2EDuration="3.296735142s" podCreationTimestamp="2026-01-22 11:45:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:45:26.293334594 +0000 UTC m=+300.138405674" watchObservedRunningTime="2026-01-22 11:45:26.296735142 +0000 UTC m=+300.141806212" Jan 22 11:45:26 crc kubenswrapper[4874]: I0122 11:45:26.340298 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" podStartSLOduration=3.340276188 podStartE2EDuration="3.340276188s" podCreationTimestamp="2026-01-22 11:45:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:45:26.337819031 +0000 UTC m=+300.182890091" watchObservedRunningTime="2026-01-22 11:45:26.340276188 +0000 UTC m=+300.185347258" Jan 22 11:45:26 crc kubenswrapper[4874]: I0122 11:45:26.552882 4874 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 22 11:45:43 crc kubenswrapper[4874]: I0122 11:45:43.428180 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7"] Jan 22 11:45:43 crc kubenswrapper[4874]: I0122 11:45:43.431100 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" podUID="9e9625e0-ef53-4afd-9576-a41ad312a97d" containerName="controller-manager" containerID="cri-o://5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983" gracePeriod=30 Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.048753 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.200832 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-proxy-ca-bundles\") pod \"9e9625e0-ef53-4afd-9576-a41ad312a97d\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.200925 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-config\") pod \"9e9625e0-ef53-4afd-9576-a41ad312a97d\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.200983 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e9625e0-ef53-4afd-9576-a41ad312a97d-serving-cert\") pod \"9e9625e0-ef53-4afd-9576-a41ad312a97d\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.201122 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rmmk\" (UniqueName: \"kubernetes.io/projected/9e9625e0-ef53-4afd-9576-a41ad312a97d-kube-api-access-2rmmk\") pod \"9e9625e0-ef53-4afd-9576-a41ad312a97d\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.201185 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-client-ca\") pod \"9e9625e0-ef53-4afd-9576-a41ad312a97d\" (UID: \"9e9625e0-ef53-4afd-9576-a41ad312a97d\") " Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.202025 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9e9625e0-ef53-4afd-9576-a41ad312a97d" (UID: "9e9625e0-ef53-4afd-9576-a41ad312a97d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.202163 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-client-ca" (OuterVolumeSpecName: "client-ca") pod "9e9625e0-ef53-4afd-9576-a41ad312a97d" (UID: "9e9625e0-ef53-4afd-9576-a41ad312a97d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.202230 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-config" (OuterVolumeSpecName: "config") pod "9e9625e0-ef53-4afd-9576-a41ad312a97d" (UID: "9e9625e0-ef53-4afd-9576-a41ad312a97d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.202831 4874 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.202881 4874 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.202909 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9625e0-ef53-4afd-9576-a41ad312a97d-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.210526 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9625e0-ef53-4afd-9576-a41ad312a97d-kube-api-access-2rmmk" (OuterVolumeSpecName: "kube-api-access-2rmmk") pod "9e9625e0-ef53-4afd-9576-a41ad312a97d" (UID: "9e9625e0-ef53-4afd-9576-a41ad312a97d"). InnerVolumeSpecName "kube-api-access-2rmmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.214137 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e9625e0-ef53-4afd-9576-a41ad312a97d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9e9625e0-ef53-4afd-9576-a41ad312a97d" (UID: "9e9625e0-ef53-4afd-9576-a41ad312a97d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.303460 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e9625e0-ef53-4afd-9576-a41ad312a97d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.303523 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rmmk\" (UniqueName: \"kubernetes.io/projected/9e9625e0-ef53-4afd-9576-a41ad312a97d-kube-api-access-2rmmk\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.382545 4874 generic.go:334] "Generic (PLEG): container finished" podID="9e9625e0-ef53-4afd-9576-a41ad312a97d" containerID="5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983" exitCode=0 Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.382599 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" event={"ID":"9e9625e0-ef53-4afd-9576-a41ad312a97d","Type":"ContainerDied","Data":"5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983"} Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.382640 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" event={"ID":"9e9625e0-ef53-4afd-9576-a41ad312a97d","Type":"ContainerDied","Data":"d2b018a92f66b0200304d7741f10b85f1e493ae49ca6a4372b8a9605031feb95"} Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.382653 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.382664 4874 scope.go:117] "RemoveContainer" containerID="5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.408332 4874 scope.go:117] "RemoveContainer" containerID="5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983" Jan 22 11:45:44 crc kubenswrapper[4874]: E0122 11:45:44.409565 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983\": container with ID starting with 5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983 not found: ID does not exist" containerID="5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.409603 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983"} err="failed to get container status \"5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983\": rpc error: code = NotFound desc = could not find container \"5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983\": container with ID starting with 5111f552e769bb7d8a02337c79b11097d1e325a76160fdcaa161f52d7010f983 not found: ID does not exist" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.437481 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7"] Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.440513 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-789bbbcf9f-jm8f7"] Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.722781 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e9625e0-ef53-4afd-9576-a41ad312a97d" path="/var/lib/kubelet/pods/9e9625e0-ef53-4afd-9576-a41ad312a97d/volumes" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.734281 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-869779d669-w4m58"] Jan 22 11:45:44 crc kubenswrapper[4874]: E0122 11:45:44.734490 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9625e0-ef53-4afd-9576-a41ad312a97d" containerName="controller-manager" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.734501 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9625e0-ef53-4afd-9576-a41ad312a97d" containerName="controller-manager" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.734606 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9625e0-ef53-4afd-9576-a41ad312a97d" containerName="controller-manager" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.734976 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.737379 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.737796 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.737988 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.738372 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.739204 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.741221 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.749619 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.755739 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-869779d669-w4m58"] Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.809520 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7305b4b5-c89c-4134-a5bf-f2e001268cb9-proxy-ca-bundles\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.809565 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4p2h\" (UniqueName: \"kubernetes.io/projected/7305b4b5-c89c-4134-a5bf-f2e001268cb9-kube-api-access-g4p2h\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.809594 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7305b4b5-c89c-4134-a5bf-f2e001268cb9-client-ca\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.809627 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7305b4b5-c89c-4134-a5bf-f2e001268cb9-serving-cert\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.809645 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7305b4b5-c89c-4134-a5bf-f2e001268cb9-config\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.910358 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7305b4b5-c89c-4134-a5bf-f2e001268cb9-serving-cert\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.910422 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7305b4b5-c89c-4134-a5bf-f2e001268cb9-config\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.910480 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7305b4b5-c89c-4134-a5bf-f2e001268cb9-proxy-ca-bundles\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.910506 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4p2h\" (UniqueName: \"kubernetes.io/projected/7305b4b5-c89c-4134-a5bf-f2e001268cb9-kube-api-access-g4p2h\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.910534 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7305b4b5-c89c-4134-a5bf-f2e001268cb9-client-ca\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.911558 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7305b4b5-c89c-4134-a5bf-f2e001268cb9-client-ca\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.912164 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7305b4b5-c89c-4134-a5bf-f2e001268cb9-proxy-ca-bundles\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.912809 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7305b4b5-c89c-4134-a5bf-f2e001268cb9-config\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.915311 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7305b4b5-c89c-4134-a5bf-f2e001268cb9-serving-cert\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:44 crc kubenswrapper[4874]: I0122 11:45:44.937515 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4p2h\" (UniqueName: \"kubernetes.io/projected/7305b4b5-c89c-4134-a5bf-f2e001268cb9-kube-api-access-g4p2h\") pod \"controller-manager-869779d669-w4m58\" (UID: \"7305b4b5-c89c-4134-a5bf-f2e001268cb9\") " pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:45 crc kubenswrapper[4874]: I0122 11:45:45.095656 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:45 crc kubenswrapper[4874]: I0122 11:45:45.303856 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-869779d669-w4m58"] Jan 22 11:45:45 crc kubenswrapper[4874]: W0122 11:45:45.326853 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7305b4b5_c89c_4134_a5bf_f2e001268cb9.slice/crio-e7303afdd5908ee0cf294c27cca6bd62b544b77d9d7c8178430ca1b757a66a87 WatchSource:0}: Error finding container e7303afdd5908ee0cf294c27cca6bd62b544b77d9d7c8178430ca1b757a66a87: Status 404 returned error can't find the container with id e7303afdd5908ee0cf294c27cca6bd62b544b77d9d7c8178430ca1b757a66a87 Jan 22 11:45:45 crc kubenswrapper[4874]: I0122 11:45:45.397331 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-869779d669-w4m58" event={"ID":"7305b4b5-c89c-4134-a5bf-f2e001268cb9","Type":"ContainerStarted","Data":"e7303afdd5908ee0cf294c27cca6bd62b544b77d9d7c8178430ca1b757a66a87"} Jan 22 11:45:46 crc kubenswrapper[4874]: I0122 11:45:46.404727 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-869779d669-w4m58" event={"ID":"7305b4b5-c89c-4134-a5bf-f2e001268cb9","Type":"ContainerStarted","Data":"8fe41fa666f6946fe9c91397a097869668c95495b0b14206c9f71d56bac2e3c8"} Jan 22 11:45:46 crc kubenswrapper[4874]: I0122 11:45:46.405082 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:46 crc kubenswrapper[4874]: I0122 11:45:46.409278 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-869779d669-w4m58" Jan 22 11:45:46 crc kubenswrapper[4874]: I0122 11:45:46.423348 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-869779d669-w4m58" podStartSLOduration=3.423330621 podStartE2EDuration="3.423330621s" podCreationTimestamp="2026-01-22 11:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:45:46.422881647 +0000 UTC m=+320.267952757" watchObservedRunningTime="2026-01-22 11:45:46.423330621 +0000 UTC m=+320.268401691" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.380552 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r2mwt"] Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.381178 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bk5qd"] Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.381492 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bk5qd" podUID="829e346d-eb89-4705-83c4-99d02fca8971" containerName="registry-server" containerID="cri-o://051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e" gracePeriod=30 Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.381764 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r2mwt" podUID="dea4a6eb-c0b1-432a-81f2-e417250b0138" containerName="registry-server" containerID="cri-o://1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32" gracePeriod=30 Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.382899 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6dvvn"] Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.383170 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" podUID="213d34f5-75cd-459c-9e56-2938fe5e3950" containerName="marketplace-operator" containerID="cri-o://a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb" gracePeriod=30 Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.406847 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhb8k"] Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.407083 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xhb8k" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerName="registry-server" containerID="cri-o://c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a" gracePeriod=30 Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.414926 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s2gqx"] Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.415145 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s2gqx" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerName="registry-server" containerID="cri-o://d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd" gracePeriod=30 Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.424743 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-st8lg"] Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.425675 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.425721 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-st8lg"] Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.492026 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-n2lnd"] Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.492662 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.526191 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-n2lnd"] Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.566721 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtdms\" (UniqueName: \"kubernetes.io/projected/bb25a3d3-60b1-43ae-b007-19b20c362414-kube-api-access-dtdms\") pod \"marketplace-operator-79b997595-st8lg\" (UID: \"bb25a3d3-60b1-43ae-b007-19b20c362414\") " pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.566774 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb25a3d3-60b1-43ae-b007-19b20c362414-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-st8lg\" (UID: \"bb25a3d3-60b1-43ae-b007-19b20c362414\") " pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.566808 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb25a3d3-60b1-43ae-b007-19b20c362414-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-st8lg\" (UID: \"bb25a3d3-60b1-43ae-b007-19b20c362414\") " pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.668554 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-trusted-ca\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.668607 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.668787 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtdms\" (UniqueName: \"kubernetes.io/projected/bb25a3d3-60b1-43ae-b007-19b20c362414-kube-api-access-dtdms\") pod \"marketplace-operator-79b997595-st8lg\" (UID: \"bb25a3d3-60b1-43ae-b007-19b20c362414\") " pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.668831 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-bound-sa-token\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.668871 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb25a3d3-60b1-43ae-b007-19b20c362414-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-st8lg\" (UID: \"bb25a3d3-60b1-43ae-b007-19b20c362414\") " pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.668896 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-registry-certificates\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.668929 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb25a3d3-60b1-43ae-b007-19b20c362414-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-st8lg\" (UID: \"bb25a3d3-60b1-43ae-b007-19b20c362414\") " pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.668983 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.669021 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-registry-tls\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.669036 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgfrm\" (UniqueName: \"kubernetes.io/projected/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-kube-api-access-rgfrm\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.669057 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.670974 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb25a3d3-60b1-43ae-b007-19b20c362414-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-st8lg\" (UID: \"bb25a3d3-60b1-43ae-b007-19b20c362414\") " pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.674626 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb25a3d3-60b1-43ae-b007-19b20c362414-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-st8lg\" (UID: \"bb25a3d3-60b1-43ae-b007-19b20c362414\") " pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.686129 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtdms\" (UniqueName: \"kubernetes.io/projected/bb25a3d3-60b1-43ae-b007-19b20c362414-kube-api-access-dtdms\") pod \"marketplace-operator-79b997595-st8lg\" (UID: \"bb25a3d3-60b1-43ae-b007-19b20c362414\") " pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.692330 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.742716 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.771008 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-registry-certificates\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.771078 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.771118 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-registry-tls\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.771152 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgfrm\" (UniqueName: \"kubernetes.io/projected/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-kube-api-access-rgfrm\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.771191 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-trusted-ca\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.771263 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.771326 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-bound-sa-token\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.774143 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.775816 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.778603 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-registry-tls\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.781328 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-trusted-ca\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.788589 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-registry-certificates\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.792257 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-bound-sa-token\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.796802 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgfrm\" (UniqueName: \"kubernetes.io/projected/c6b1b602-57a0-4459-bb14-e96f6fa1fe9c-kube-api-access-rgfrm\") pod \"image-registry-66df7c8f76-n2lnd\" (UID: \"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.827263 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:49 crc kubenswrapper[4874]: I0122 11:45:49.909679 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.067147 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.078480 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9tqj\" (UniqueName: \"kubernetes.io/projected/829e346d-eb89-4705-83c4-99d02fca8971-kube-api-access-k9tqj\") pod \"829e346d-eb89-4705-83c4-99d02fca8971\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.078600 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-catalog-content\") pod \"829e346d-eb89-4705-83c4-99d02fca8971\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.078657 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-utilities\") pod \"829e346d-eb89-4705-83c4-99d02fca8971\" (UID: \"829e346d-eb89-4705-83c4-99d02fca8971\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.079860 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-utilities" (OuterVolumeSpecName: "utilities") pod "829e346d-eb89-4705-83c4-99d02fca8971" (UID: "829e346d-eb89-4705-83c4-99d02fca8971"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.090054 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829e346d-eb89-4705-83c4-99d02fca8971-kube-api-access-k9tqj" (OuterVolumeSpecName: "kube-api-access-k9tqj") pod "829e346d-eb89-4705-83c4-99d02fca8971" (UID: "829e346d-eb89-4705-83c4-99d02fca8971"). InnerVolumeSpecName "kube-api-access-k9tqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.101195 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.102818 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.122448 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.162856 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "829e346d-eb89-4705-83c4-99d02fca8971" (UID: "829e346d-eb89-4705-83c4-99d02fca8971"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.179550 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79nw7\" (UniqueName: \"kubernetes.io/projected/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-kube-api-access-79nw7\") pod \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.179616 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-utilities\") pod \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.179769 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-catalog-content\") pod \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\" (UID: \"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.180117 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.180151 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9tqj\" (UniqueName: \"kubernetes.io/projected/829e346d-eb89-4705-83c4-99d02fca8971-kube-api-access-k9tqj\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.180164 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/829e346d-eb89-4705-83c4-99d02fca8971-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.180966 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-utilities" (OuterVolumeSpecName: "utilities") pod "09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" (UID: "09a47d07-9bf6-4033-8c08-cc3aef9fe4f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.181823 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-kube-api-access-79nw7" (OuterVolumeSpecName: "kube-api-access-79nw7") pod "09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" (UID: "09a47d07-9bf6-4033-8c08-cc3aef9fe4f4"). InnerVolumeSpecName "kube-api-access-79nw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.207106 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" (UID: "09a47d07-9bf6-4033-8c08-cc3aef9fe4f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.281439 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-trusted-ca\") pod \"213d34f5-75cd-459c-9e56-2938fe5e3950\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.281558 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-catalog-content\") pod \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.281639 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-operator-metrics\") pod \"213d34f5-75cd-459c-9e56-2938fe5e3950\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.281689 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2xvq\" (UniqueName: \"kubernetes.io/projected/dea4a6eb-c0b1-432a-81f2-e417250b0138-kube-api-access-w2xvq\") pod \"dea4a6eb-c0b1-432a-81f2-e417250b0138\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.281747 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-catalog-content\") pod \"dea4a6eb-c0b1-432a-81f2-e417250b0138\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.281771 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-utilities\") pod \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.281803 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf6nt\" (UniqueName: \"kubernetes.io/projected/213d34f5-75cd-459c-9e56-2938fe5e3950-kube-api-access-lf6nt\") pod \"213d34f5-75cd-459c-9e56-2938fe5e3950\" (UID: \"213d34f5-75cd-459c-9e56-2938fe5e3950\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.281858 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92txx\" (UniqueName: \"kubernetes.io/projected/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-kube-api-access-92txx\") pod \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\" (UID: \"2d07947b-508d-4f12-ba1b-2d5f24a6db2c\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.282541 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-utilities" (OuterVolumeSpecName: "utilities") pod "2d07947b-508d-4f12-ba1b-2d5f24a6db2c" (UID: "2d07947b-508d-4f12-ba1b-2d5f24a6db2c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.282568 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "213d34f5-75cd-459c-9e56-2938fe5e3950" (UID: "213d34f5-75cd-459c-9e56-2938fe5e3950"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.282732 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-utilities\") pod \"dea4a6eb-c0b1-432a-81f2-e417250b0138\" (UID: \"dea4a6eb-c0b1-432a-81f2-e417250b0138\") " Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.283948 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-utilities" (OuterVolumeSpecName: "utilities") pod "dea4a6eb-c0b1-432a-81f2-e417250b0138" (UID: "dea4a6eb-c0b1-432a-81f2-e417250b0138"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.284683 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.284703 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79nw7\" (UniqueName: \"kubernetes.io/projected/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-kube-api-access-79nw7\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.284717 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.284729 4874 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.284763 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.284774 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.285908 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-kube-api-access-92txx" (OuterVolumeSpecName: "kube-api-access-92txx") pod "2d07947b-508d-4f12-ba1b-2d5f24a6db2c" (UID: "2d07947b-508d-4f12-ba1b-2d5f24a6db2c"). InnerVolumeSpecName "kube-api-access-92txx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.285784 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dea4a6eb-c0b1-432a-81f2-e417250b0138-kube-api-access-w2xvq" (OuterVolumeSpecName: "kube-api-access-w2xvq") pod "dea4a6eb-c0b1-432a-81f2-e417250b0138" (UID: "dea4a6eb-c0b1-432a-81f2-e417250b0138"). InnerVolumeSpecName "kube-api-access-w2xvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.286057 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213d34f5-75cd-459c-9e56-2938fe5e3950-kube-api-access-lf6nt" (OuterVolumeSpecName: "kube-api-access-lf6nt") pod "213d34f5-75cd-459c-9e56-2938fe5e3950" (UID: "213d34f5-75cd-459c-9e56-2938fe5e3950"). InnerVolumeSpecName "kube-api-access-lf6nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.289432 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "213d34f5-75cd-459c-9e56-2938fe5e3950" (UID: "213d34f5-75cd-459c-9e56-2938fe5e3950"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.315935 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-st8lg"] Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.334651 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dea4a6eb-c0b1-432a-81f2-e417250b0138" (UID: "dea4a6eb-c0b1-432a-81f2-e417250b0138"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.385886 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dea4a6eb-c0b1-432a-81f2-e417250b0138-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.386220 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf6nt\" (UniqueName: \"kubernetes.io/projected/213d34f5-75cd-459c-9e56-2938fe5e3950-kube-api-access-lf6nt\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.386234 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92txx\" (UniqueName: \"kubernetes.io/projected/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-kube-api-access-92txx\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.386275 4874 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/213d34f5-75cd-459c-9e56-2938fe5e3950-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.386291 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2xvq\" (UniqueName: \"kubernetes.io/projected/dea4a6eb-c0b1-432a-81f2-e417250b0138-kube-api-access-w2xvq\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.412677 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d07947b-508d-4f12-ba1b-2d5f24a6db2c" (UID: "2d07947b-508d-4f12-ba1b-2d5f24a6db2c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.415906 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-n2lnd"] Jan 22 11:45:50 crc kubenswrapper[4874]: W0122 11:45:50.420843 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6b1b602_57a0_4459_bb14_e96f6fa1fe9c.slice/crio-ed2fc786963a7134e138b006ccc00a8ca69ad6f8e26c223c0b234675e09772cf WatchSource:0}: Error finding container ed2fc786963a7134e138b006ccc00a8ca69ad6f8e26c223c0b234675e09772cf: Status 404 returned error can't find the container with id ed2fc786963a7134e138b006ccc00a8ca69ad6f8e26c223c0b234675e09772cf Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.451895 4874 generic.go:334] "Generic (PLEG): container finished" podID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerID="c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a" exitCode=0 Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.451983 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhb8k" event={"ID":"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4","Type":"ContainerDied","Data":"c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a"} Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.452025 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhb8k" event={"ID":"09a47d07-9bf6-4033-8c08-cc3aef9fe4f4","Type":"ContainerDied","Data":"4f0dfc138b95fd589daf732c95563b35b123c0075950db736b607653c673e3e1"} Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.452042 4874 scope.go:117] "RemoveContainer" containerID="c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.452216 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhb8k" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.456476 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" event={"ID":"bb25a3d3-60b1-43ae-b007-19b20c362414","Type":"ContainerStarted","Data":"86211cce9bb8758fe37c94e0307689a34a40a3b0ea5d57981af9aeb1748168f6"} Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.458476 4874 generic.go:334] "Generic (PLEG): container finished" podID="213d34f5-75cd-459c-9e56-2938fe5e3950" containerID="a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb" exitCode=0 Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.458516 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" event={"ID":"213d34f5-75cd-459c-9e56-2938fe5e3950","Type":"ContainerDied","Data":"a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb"} Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.458544 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" event={"ID":"213d34f5-75cd-459c-9e56-2938fe5e3950","Type":"ContainerDied","Data":"e55fae314e5ccbaff31d4019382891789d84d2a8792b6ac6cab4bdcc20e97e1e"} Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.458626 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6dvvn" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.467845 4874 generic.go:334] "Generic (PLEG): container finished" podID="829e346d-eb89-4705-83c4-99d02fca8971" containerID="051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e" exitCode=0 Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.467923 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bk5qd" event={"ID":"829e346d-eb89-4705-83c4-99d02fca8971","Type":"ContainerDied","Data":"051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e"} Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.467950 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bk5qd" event={"ID":"829e346d-eb89-4705-83c4-99d02fca8971","Type":"ContainerDied","Data":"06e653a8a66624e58a13fce521592f82bf5da1e8eff383092cdaf8dbc041daca"} Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.467957 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bk5qd" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.478695 4874 generic.go:334] "Generic (PLEG): container finished" podID="dea4a6eb-c0b1-432a-81f2-e417250b0138" containerID="1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32" exitCode=0 Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.478766 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2mwt" event={"ID":"dea4a6eb-c0b1-432a-81f2-e417250b0138","Type":"ContainerDied","Data":"1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32"} Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.478793 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2mwt" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.478806 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2mwt" event={"ID":"dea4a6eb-c0b1-432a-81f2-e417250b0138","Type":"ContainerDied","Data":"44f0f27f3f6817f2e87df4da350d3341a31946833627d7893617f4278baf12f8"} Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.491809 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d07947b-508d-4f12-ba1b-2d5f24a6db2c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.495332 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhb8k"] Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.502911 4874 scope.go:117] "RemoveContainer" containerID="0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.507191 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhb8k"] Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.511680 4874 generic.go:334] "Generic (PLEG): container finished" podID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerID="d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd" exitCode=0 Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.511746 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2gqx" event={"ID":"2d07947b-508d-4f12-ba1b-2d5f24a6db2c","Type":"ContainerDied","Data":"d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd"} Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.511794 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2gqx" event={"ID":"2d07947b-508d-4f12-ba1b-2d5f24a6db2c","Type":"ContainerDied","Data":"880843b24ca67a3147842f7c7ec48125fab7fce5de031516ba04e79aff10f06a"} Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.511919 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2gqx" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.564496 4874 scope.go:117] "RemoveContainer" containerID="af7c8535e94f5d42ca318e78dacd324021ec14636cac5b8d6617938e7cede85c" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.606130 4874 scope.go:117] "RemoveContainer" containerID="c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.606708 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a\": container with ID starting with c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a not found: ID does not exist" containerID="c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.606745 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a"} err="failed to get container status \"c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a\": rpc error: code = NotFound desc = could not find container \"c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a\": container with ID starting with c92c089eac1a4f555258a56bb484b8c71a6cc82dce9149ac96cfba7e4b8f946a not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.606772 4874 scope.go:117] "RemoveContainer" containerID="0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.607225 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f\": container with ID starting with 0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f not found: ID does not exist" containerID="0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.607266 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f"} err="failed to get container status \"0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f\": rpc error: code = NotFound desc = could not find container \"0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f\": container with ID starting with 0271a88b4f392b5c007a29012fdaee2d611a90d52f146f49b1b948112a16ea7f not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.607292 4874 scope.go:117] "RemoveContainer" containerID="af7c8535e94f5d42ca318e78dacd324021ec14636cac5b8d6617938e7cede85c" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.607675 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af7c8535e94f5d42ca318e78dacd324021ec14636cac5b8d6617938e7cede85c\": container with ID starting with af7c8535e94f5d42ca318e78dacd324021ec14636cac5b8d6617938e7cede85c not found: ID does not exist" containerID="af7c8535e94f5d42ca318e78dacd324021ec14636cac5b8d6617938e7cede85c" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.607703 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af7c8535e94f5d42ca318e78dacd324021ec14636cac5b8d6617938e7cede85c"} err="failed to get container status \"af7c8535e94f5d42ca318e78dacd324021ec14636cac5b8d6617938e7cede85c\": rpc error: code = NotFound desc = could not find container \"af7c8535e94f5d42ca318e78dacd324021ec14636cac5b8d6617938e7cede85c\": container with ID starting with af7c8535e94f5d42ca318e78dacd324021ec14636cac5b8d6617938e7cede85c not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.607720 4874 scope.go:117] "RemoveContainer" containerID="a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.628661 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bk5qd"] Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.629627 4874 scope.go:117] "RemoveContainer" containerID="a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.634541 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bk5qd"] Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.646825 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6dvvn"] Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.648987 4874 scope.go:117] "RemoveContainer" containerID="a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.649586 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb\": container with ID starting with a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb not found: ID does not exist" containerID="a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.649787 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb"} err="failed to get container status \"a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb\": rpc error: code = NotFound desc = could not find container \"a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb\": container with ID starting with a89cf0f40681e6ea2394a8efa5206b34afe35c8149f80b74967ca70f66ececcb not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.649878 4874 scope.go:117] "RemoveContainer" containerID="a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.650005 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6dvvn"] Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.650212 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3\": container with ID starting with a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3 not found: ID does not exist" containerID="a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.650257 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3"} err="failed to get container status \"a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3\": rpc error: code = NotFound desc = could not find container \"a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3\": container with ID starting with a3ab1ecac652d1663a1d25c20a4f94b60f4dcf52866d48a08ff55d35fc806fe3 not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.650277 4874 scope.go:117] "RemoveContainer" containerID="051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.660480 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r2mwt"] Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.664004 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r2mwt"] Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.667053 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s2gqx"] Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.669645 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s2gqx"] Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.670769 4874 scope.go:117] "RemoveContainer" containerID="231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.685699 4874 scope.go:117] "RemoveContainer" containerID="e04166d257b1ab6ec62d2fb4997b1ee33b374447294ba53534e3f997572dbcbc" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.698422 4874 scope.go:117] "RemoveContainer" containerID="051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.700259 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e\": container with ID starting with 051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e not found: ID does not exist" containerID="051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.700317 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e"} err="failed to get container status \"051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e\": rpc error: code = NotFound desc = could not find container \"051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e\": container with ID starting with 051c0e6931085c66795c18888f3c00c7f7391d31df635135b32e59153b20723e not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.700362 4874 scope.go:117] "RemoveContainer" containerID="231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.700981 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b\": container with ID starting with 231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b not found: ID does not exist" containerID="231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.701014 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b"} err="failed to get container status \"231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b\": rpc error: code = NotFound desc = could not find container \"231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b\": container with ID starting with 231aba9fc47370e4a6cd041c18aeb432ac088da84612e9ae338664bb8b47b53b not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.701040 4874 scope.go:117] "RemoveContainer" containerID="e04166d257b1ab6ec62d2fb4997b1ee33b374447294ba53534e3f997572dbcbc" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.701235 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e04166d257b1ab6ec62d2fb4997b1ee33b374447294ba53534e3f997572dbcbc\": container with ID starting with e04166d257b1ab6ec62d2fb4997b1ee33b374447294ba53534e3f997572dbcbc not found: ID does not exist" containerID="e04166d257b1ab6ec62d2fb4997b1ee33b374447294ba53534e3f997572dbcbc" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.701286 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e04166d257b1ab6ec62d2fb4997b1ee33b374447294ba53534e3f997572dbcbc"} err="failed to get container status \"e04166d257b1ab6ec62d2fb4997b1ee33b374447294ba53534e3f997572dbcbc\": rpc error: code = NotFound desc = could not find container \"e04166d257b1ab6ec62d2fb4997b1ee33b374447294ba53534e3f997572dbcbc\": container with ID starting with e04166d257b1ab6ec62d2fb4997b1ee33b374447294ba53534e3f997572dbcbc not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.701303 4874 scope.go:117] "RemoveContainer" containerID="1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.713140 4874 scope.go:117] "RemoveContainer" containerID="3afe6d67a395d98d3557fc317e71f0c8a6bb5a855419905a6dab04529d3efc0a" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.723961 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" path="/var/lib/kubelet/pods/09a47d07-9bf6-4033-8c08-cc3aef9fe4f4/volumes" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.724825 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="213d34f5-75cd-459c-9e56-2938fe5e3950" path="/var/lib/kubelet/pods/213d34f5-75cd-459c-9e56-2938fe5e3950/volumes" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.725385 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" path="/var/lib/kubelet/pods/2d07947b-508d-4f12-ba1b-2d5f24a6db2c/volumes" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.726668 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="829e346d-eb89-4705-83c4-99d02fca8971" path="/var/lib/kubelet/pods/829e346d-eb89-4705-83c4-99d02fca8971/volumes" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.726907 4874 scope.go:117] "RemoveContainer" containerID="f102fff60f0bad6a90aabfcc1da36f807499c6f5b663e355666a3b9418698f61" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.727388 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dea4a6eb-c0b1-432a-81f2-e417250b0138" path="/var/lib/kubelet/pods/dea4a6eb-c0b1-432a-81f2-e417250b0138/volumes" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.737369 4874 scope.go:117] "RemoveContainer" containerID="1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.740867 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32\": container with ID starting with 1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32 not found: ID does not exist" containerID="1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.740895 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32"} err="failed to get container status \"1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32\": rpc error: code = NotFound desc = could not find container \"1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32\": container with ID starting with 1f66114e745014e9b2811f8ce4a3519c0db97b1d4c5d32326d52ab45a752ab32 not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.740915 4874 scope.go:117] "RemoveContainer" containerID="3afe6d67a395d98d3557fc317e71f0c8a6bb5a855419905a6dab04529d3efc0a" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.741159 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3afe6d67a395d98d3557fc317e71f0c8a6bb5a855419905a6dab04529d3efc0a\": container with ID starting with 3afe6d67a395d98d3557fc317e71f0c8a6bb5a855419905a6dab04529d3efc0a not found: ID does not exist" containerID="3afe6d67a395d98d3557fc317e71f0c8a6bb5a855419905a6dab04529d3efc0a" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.741178 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afe6d67a395d98d3557fc317e71f0c8a6bb5a855419905a6dab04529d3efc0a"} err="failed to get container status \"3afe6d67a395d98d3557fc317e71f0c8a6bb5a855419905a6dab04529d3efc0a\": rpc error: code = NotFound desc = could not find container \"3afe6d67a395d98d3557fc317e71f0c8a6bb5a855419905a6dab04529d3efc0a\": container with ID starting with 3afe6d67a395d98d3557fc317e71f0c8a6bb5a855419905a6dab04529d3efc0a not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.741191 4874 scope.go:117] "RemoveContainer" containerID="f102fff60f0bad6a90aabfcc1da36f807499c6f5b663e355666a3b9418698f61" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.741440 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f102fff60f0bad6a90aabfcc1da36f807499c6f5b663e355666a3b9418698f61\": container with ID starting with f102fff60f0bad6a90aabfcc1da36f807499c6f5b663e355666a3b9418698f61 not found: ID does not exist" containerID="f102fff60f0bad6a90aabfcc1da36f807499c6f5b663e355666a3b9418698f61" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.741461 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f102fff60f0bad6a90aabfcc1da36f807499c6f5b663e355666a3b9418698f61"} err="failed to get container status \"f102fff60f0bad6a90aabfcc1da36f807499c6f5b663e355666a3b9418698f61\": rpc error: code = NotFound desc = could not find container \"f102fff60f0bad6a90aabfcc1da36f807499c6f5b663e355666a3b9418698f61\": container with ID starting with f102fff60f0bad6a90aabfcc1da36f807499c6f5b663e355666a3b9418698f61 not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.741473 4874 scope.go:117] "RemoveContainer" containerID="d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.762862 4874 scope.go:117] "RemoveContainer" containerID="b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.777544 4874 scope.go:117] "RemoveContainer" containerID="b45c0a0edb45e712987d45512f5f8181affde43a34f168d3078df992db0dd1ef" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.791220 4874 scope.go:117] "RemoveContainer" containerID="d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.791640 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd\": container with ID starting with d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd not found: ID does not exist" containerID="d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.791681 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd"} err="failed to get container status \"d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd\": rpc error: code = NotFound desc = could not find container \"d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd\": container with ID starting with d86e34d52dc8fcbe07403b1086c0ea0e849c1283e037e58504051c2dba7b2efd not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.791710 4874 scope.go:117] "RemoveContainer" containerID="b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.792095 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec\": container with ID starting with b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec not found: ID does not exist" containerID="b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.792125 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec"} err="failed to get container status \"b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec\": rpc error: code = NotFound desc = could not find container \"b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec\": container with ID starting with b877e60b7da09c280a2886dd4dcd5b096a9cfc5cccc418b32f7e1f7a67f9a9ec not found: ID does not exist" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.792149 4874 scope.go:117] "RemoveContainer" containerID="b45c0a0edb45e712987d45512f5f8181affde43a34f168d3078df992db0dd1ef" Jan 22 11:45:50 crc kubenswrapper[4874]: E0122 11:45:50.792442 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b45c0a0edb45e712987d45512f5f8181affde43a34f168d3078df992db0dd1ef\": container with ID starting with b45c0a0edb45e712987d45512f5f8181affde43a34f168d3078df992db0dd1ef not found: ID does not exist" containerID="b45c0a0edb45e712987d45512f5f8181affde43a34f168d3078df992db0dd1ef" Jan 22 11:45:50 crc kubenswrapper[4874]: I0122 11:45:50.792470 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45c0a0edb45e712987d45512f5f8181affde43a34f168d3078df992db0dd1ef"} err="failed to get container status \"b45c0a0edb45e712987d45512f5f8181affde43a34f168d3078df992db0dd1ef\": rpc error: code = NotFound desc = could not find container \"b45c0a0edb45e712987d45512f5f8181affde43a34f168d3078df992db0dd1ef\": container with ID starting with b45c0a0edb45e712987d45512f5f8181affde43a34f168d3078df992db0dd1ef not found: ID does not exist" Jan 22 11:45:51 crc kubenswrapper[4874]: I0122 11:45:51.519011 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" event={"ID":"bb25a3d3-60b1-43ae-b007-19b20c362414","Type":"ContainerStarted","Data":"d71bf6ae06cdecffa61111320d32d5070c4cc6a90d9ba025a3aa011af643ed86"} Jan 22 11:45:51 crc kubenswrapper[4874]: I0122 11:45:51.519273 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:51 crc kubenswrapper[4874]: I0122 11:45:51.523176 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" Jan 22 11:45:51 crc kubenswrapper[4874]: I0122 11:45:51.528182 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" event={"ID":"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c","Type":"ContainerStarted","Data":"d1f06fcfda9d856f6781243495ca08fc9b4566fd7006eab8adb7d4c9f8023bde"} Jan 22 11:45:51 crc kubenswrapper[4874]: I0122 11:45:51.528218 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" event={"ID":"c6b1b602-57a0-4459-bb14-e96f6fa1fe9c","Type":"ContainerStarted","Data":"ed2fc786963a7134e138b006ccc00a8ca69ad6f8e26c223c0b234675e09772cf"} Jan 22 11:45:51 crc kubenswrapper[4874]: I0122 11:45:51.528280 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:45:51 crc kubenswrapper[4874]: I0122 11:45:51.539209 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-st8lg" podStartSLOduration=2.539185818 podStartE2EDuration="2.539185818s" podCreationTimestamp="2026-01-22 11:45:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:45:51.536170863 +0000 UTC m=+325.381241933" watchObservedRunningTime="2026-01-22 11:45:51.539185818 +0000 UTC m=+325.384256888" Jan 22 11:45:51 crc kubenswrapper[4874]: I0122 11:45:51.572071 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" podStartSLOduration=2.57205121 podStartE2EDuration="2.57205121s" podCreationTimestamp="2026-01-22 11:45:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:45:51.568876321 +0000 UTC m=+325.413947391" watchObservedRunningTime="2026-01-22 11:45:51.57205121 +0000 UTC m=+325.417122290" Jan 22 11:46:03 crc kubenswrapper[4874]: I0122 11:46:03.372872 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2"] Jan 22 11:46:03 crc kubenswrapper[4874]: I0122 11:46:03.373546 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" podUID="8ba4583a-60a2-4ace-ba80-8d804556d604" containerName="route-controller-manager" containerID="cri-o://21043c10d05ec41a76bd77364b386c886df31d414e1e8c2604a1659796f8c27e" gracePeriod=30 Jan 22 11:46:03 crc kubenswrapper[4874]: I0122 11:46:03.590191 4874 generic.go:334] "Generic (PLEG): container finished" podID="8ba4583a-60a2-4ace-ba80-8d804556d604" containerID="21043c10d05ec41a76bd77364b386c886df31d414e1e8c2604a1659796f8c27e" exitCode=0 Jan 22 11:46:03 crc kubenswrapper[4874]: I0122 11:46:03.590233 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" event={"ID":"8ba4583a-60a2-4ace-ba80-8d804556d604","Type":"ContainerDied","Data":"21043c10d05ec41a76bd77364b386c886df31d414e1e8c2604a1659796f8c27e"} Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.308735 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.470849 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-config\") pod \"8ba4583a-60a2-4ace-ba80-8d804556d604\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.471444 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-client-ca\") pod \"8ba4583a-60a2-4ace-ba80-8d804556d604\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.471577 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xklr5\" (UniqueName: \"kubernetes.io/projected/8ba4583a-60a2-4ace-ba80-8d804556d604-kube-api-access-xklr5\") pod \"8ba4583a-60a2-4ace-ba80-8d804556d604\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.471689 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ba4583a-60a2-4ace-ba80-8d804556d604-serving-cert\") pod \"8ba4583a-60a2-4ace-ba80-8d804556d604\" (UID: \"8ba4583a-60a2-4ace-ba80-8d804556d604\") " Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.471761 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-config" (OuterVolumeSpecName: "config") pod "8ba4583a-60a2-4ace-ba80-8d804556d604" (UID: "8ba4583a-60a2-4ace-ba80-8d804556d604"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.471923 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-client-ca" (OuterVolumeSpecName: "client-ca") pod "8ba4583a-60a2-4ace-ba80-8d804556d604" (UID: "8ba4583a-60a2-4ace-ba80-8d804556d604"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.472160 4874 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.472184 4874 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ba4583a-60a2-4ace-ba80-8d804556d604-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.476466 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ba4583a-60a2-4ace-ba80-8d804556d604-kube-api-access-xklr5" (OuterVolumeSpecName: "kube-api-access-xklr5") pod "8ba4583a-60a2-4ace-ba80-8d804556d604" (UID: "8ba4583a-60a2-4ace-ba80-8d804556d604"). InnerVolumeSpecName "kube-api-access-xklr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.478466 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba4583a-60a2-4ace-ba80-8d804556d604-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8ba4583a-60a2-4ace-ba80-8d804556d604" (UID: "8ba4583a-60a2-4ace-ba80-8d804556d604"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.535988 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5s7gc"] Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536187 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829e346d-eb89-4705-83c4-99d02fca8971" containerName="extract-utilities" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536200 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="829e346d-eb89-4705-83c4-99d02fca8971" containerName="extract-utilities" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536210 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerName="extract-content" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536216 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerName="extract-content" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536225 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829e346d-eb89-4705-83c4-99d02fca8971" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536232 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="829e346d-eb89-4705-83c4-99d02fca8971" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536239 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536245 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536255 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerName="extract-content" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536260 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerName="extract-content" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536270 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea4a6eb-c0b1-432a-81f2-e417250b0138" containerName="extract-utilities" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536275 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea4a6eb-c0b1-432a-81f2-e417250b0138" containerName="extract-utilities" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536282 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829e346d-eb89-4705-83c4-99d02fca8971" containerName="extract-content" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536289 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="829e346d-eb89-4705-83c4-99d02fca8971" containerName="extract-content" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536296 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerName="extract-utilities" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536301 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerName="extract-utilities" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536308 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ba4583a-60a2-4ace-ba80-8d804556d604" containerName="route-controller-manager" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536314 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ba4583a-60a2-4ace-ba80-8d804556d604" containerName="route-controller-manager" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536323 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213d34f5-75cd-459c-9e56-2938fe5e3950" containerName="marketplace-operator" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536328 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="213d34f5-75cd-459c-9e56-2938fe5e3950" containerName="marketplace-operator" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536335 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea4a6eb-c0b1-432a-81f2-e417250b0138" containerName="extract-content" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536342 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea4a6eb-c0b1-432a-81f2-e417250b0138" containerName="extract-content" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536353 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea4a6eb-c0b1-432a-81f2-e417250b0138" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536359 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea4a6eb-c0b1-432a-81f2-e417250b0138" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536366 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536373 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536381 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerName="extract-utilities" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536387 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerName="extract-utilities" Jan 22 11:46:04 crc kubenswrapper[4874]: E0122 11:46:04.536423 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213d34f5-75cd-459c-9e56-2938fe5e3950" containerName="marketplace-operator" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536430 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="213d34f5-75cd-459c-9e56-2938fe5e3950" containerName="marketplace-operator" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536525 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="213d34f5-75cd-459c-9e56-2938fe5e3950" containerName="marketplace-operator" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536536 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="213d34f5-75cd-459c-9e56-2938fe5e3950" containerName="marketplace-operator" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536546 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="829e346d-eb89-4705-83c4-99d02fca8971" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536553 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea4a6eb-c0b1-432a-81f2-e417250b0138" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536560 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d07947b-508d-4f12-ba1b-2d5f24a6db2c" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536567 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ba4583a-60a2-4ace-ba80-8d804556d604" containerName="route-controller-manager" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.536574 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="09a47d07-9bf6-4033-8c08-cc3aef9fe4f4" containerName="registry-server" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.537198 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.540467 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.546260 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s7gc"] Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.573641 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xklr5\" (UniqueName: \"kubernetes.io/projected/8ba4583a-60a2-4ace-ba80-8d804556d604-kube-api-access-xklr5\") on node \"crc\" DevicePath \"\"" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.573680 4874 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ba4583a-60a2-4ace-ba80-8d804556d604-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.596155 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" event={"ID":"8ba4583a-60a2-4ace-ba80-8d804556d604","Type":"ContainerDied","Data":"ef570dea7a83f0ce123b85b79bfbc070103a700770f9817e620b420f82ef7f8a"} Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.596198 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.596207 4874 scope.go:117] "RemoveContainer" containerID="21043c10d05ec41a76bd77364b386c886df31d414e1e8c2604a1659796f8c27e" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.622327 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2"] Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.625901 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5764494d47-vsfq2"] Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.674830 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-catalog-content\") pod \"redhat-marketplace-5s7gc\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.674900 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-utilities\") pod \"redhat-marketplace-5s7gc\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.674928 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slv5g\" (UniqueName: \"kubernetes.io/projected/37153c5d-6533-4973-99bd-f682b3c148d9-kube-api-access-slv5g\") pod \"redhat-marketplace-5s7gc\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.725965 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ba4583a-60a2-4ace-ba80-8d804556d604" path="/var/lib/kubelet/pods/8ba4583a-60a2-4ace-ba80-8d804556d604/volumes" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.755219 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n"] Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.755981 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.757734 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.758002 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.759202 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.759454 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.759616 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.759643 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.764277 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n"] Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.775945 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-catalog-content\") pod \"redhat-marketplace-5s7gc\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.776020 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-utilities\") pod \"redhat-marketplace-5s7gc\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.776047 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slv5g\" (UniqueName: \"kubernetes.io/projected/37153c5d-6533-4973-99bd-f682b3c148d9-kube-api-access-slv5g\") pod \"redhat-marketplace-5s7gc\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.776644 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-catalog-content\") pod \"redhat-marketplace-5s7gc\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.776815 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-utilities\") pod \"redhat-marketplace-5s7gc\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.791436 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slv5g\" (UniqueName: \"kubernetes.io/projected/37153c5d-6533-4973-99bd-f682b3c148d9-kube-api-access-slv5g\") pod \"redhat-marketplace-5s7gc\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.861048 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.877971 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-config\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.878354 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28jn8\" (UniqueName: \"kubernetes.io/projected/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-kube-api-access-28jn8\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.878752 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-client-ca\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.878785 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-serving-cert\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.979696 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-config\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.980065 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28jn8\" (UniqueName: \"kubernetes.io/projected/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-kube-api-access-28jn8\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.980140 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-client-ca\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.980159 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-serving-cert\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.982410 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-client-ca\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.986587 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-serving-cert\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.990761 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-config\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:04 crc kubenswrapper[4874]: I0122 11:46:04.997317 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28jn8\" (UniqueName: \"kubernetes.io/projected/59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e-kube-api-access-28jn8\") pod \"route-controller-manager-5b8588574d-d6k8n\" (UID: \"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e\") " pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.072730 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.095989 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s7gc"] Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.142715 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7qhz5"] Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.143843 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.152068 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.153863 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7qhz5"] Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.283243 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0684b3da-a3b2-496b-be2e-b8e0fe8e1277-catalog-content\") pod \"redhat-operators-7qhz5\" (UID: \"0684b3da-a3b2-496b-be2e-b8e0fe8e1277\") " pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.283771 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2plf4\" (UniqueName: \"kubernetes.io/projected/0684b3da-a3b2-496b-be2e-b8e0fe8e1277-kube-api-access-2plf4\") pod \"redhat-operators-7qhz5\" (UID: \"0684b3da-a3b2-496b-be2e-b8e0fe8e1277\") " pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.283811 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0684b3da-a3b2-496b-be2e-b8e0fe8e1277-utilities\") pod \"redhat-operators-7qhz5\" (UID: \"0684b3da-a3b2-496b-be2e-b8e0fe8e1277\") " pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.384857 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2plf4\" (UniqueName: \"kubernetes.io/projected/0684b3da-a3b2-496b-be2e-b8e0fe8e1277-kube-api-access-2plf4\") pod \"redhat-operators-7qhz5\" (UID: \"0684b3da-a3b2-496b-be2e-b8e0fe8e1277\") " pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.384905 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0684b3da-a3b2-496b-be2e-b8e0fe8e1277-utilities\") pod \"redhat-operators-7qhz5\" (UID: \"0684b3da-a3b2-496b-be2e-b8e0fe8e1277\") " pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.384964 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0684b3da-a3b2-496b-be2e-b8e0fe8e1277-catalog-content\") pod \"redhat-operators-7qhz5\" (UID: \"0684b3da-a3b2-496b-be2e-b8e0fe8e1277\") " pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.385601 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0684b3da-a3b2-496b-be2e-b8e0fe8e1277-utilities\") pod \"redhat-operators-7qhz5\" (UID: \"0684b3da-a3b2-496b-be2e-b8e0fe8e1277\") " pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.385606 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0684b3da-a3b2-496b-be2e-b8e0fe8e1277-catalog-content\") pod \"redhat-operators-7qhz5\" (UID: \"0684b3da-a3b2-496b-be2e-b8e0fe8e1277\") " pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.403948 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2plf4\" (UniqueName: \"kubernetes.io/projected/0684b3da-a3b2-496b-be2e-b8e0fe8e1277-kube-api-access-2plf4\") pod \"redhat-operators-7qhz5\" (UID: \"0684b3da-a3b2-496b-be2e-b8e0fe8e1277\") " pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.470503 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n"] Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.479917 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.608335 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" event={"ID":"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e","Type":"ContainerStarted","Data":"bd1fd63555b8a029a92e0f40c158d063833d3e6e437ab5d0bf5c04496ca7930c"} Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.609993 4874 generic.go:334] "Generic (PLEG): container finished" podID="37153c5d-6533-4973-99bd-f682b3c148d9" containerID="684e002757e3dc3f4607d5fe73a314e71a31507700f72d618222be0e988a644f" exitCode=0 Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.610054 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s7gc" event={"ID":"37153c5d-6533-4973-99bd-f682b3c148d9","Type":"ContainerDied","Data":"684e002757e3dc3f4607d5fe73a314e71a31507700f72d618222be0e988a644f"} Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.610076 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s7gc" event={"ID":"37153c5d-6533-4973-99bd-f682b3c148d9","Type":"ContainerStarted","Data":"3610f727eca8b59c0ef9ba7a4755e8243ae3a7e6694ae4e499109a32fd98ae07"} Jan 22 11:46:05 crc kubenswrapper[4874]: I0122 11:46:05.895778 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7qhz5"] Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.623284 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s7gc" event={"ID":"37153c5d-6533-4973-99bd-f682b3c148d9","Type":"ContainerStarted","Data":"c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb"} Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.625032 4874 generic.go:334] "Generic (PLEG): container finished" podID="0684b3da-a3b2-496b-be2e-b8e0fe8e1277" containerID="3e075910332bf00ab03e82065b1f43ae1fdcf6117854e80acd8b69026db05443" exitCode=0 Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.625136 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qhz5" event={"ID":"0684b3da-a3b2-496b-be2e-b8e0fe8e1277","Type":"ContainerDied","Data":"3e075910332bf00ab03e82065b1f43ae1fdcf6117854e80acd8b69026db05443"} Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.625185 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qhz5" event={"ID":"0684b3da-a3b2-496b-be2e-b8e0fe8e1277","Type":"ContainerStarted","Data":"2a7a120c40ba3e5721b378f27bf0a65a238c9b96f9dc2ab6b7f3892cad117222"} Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.626369 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" event={"ID":"59b05a6e-2fa5-4bfa-b7c2-47f43f7d681e","Type":"ContainerStarted","Data":"c49a07ce825a7e06fa4caecc1f69fac8b2307a3ebcf1bbe7b914ac17f5c6dee2"} Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.626587 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.633274 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.688326 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5b8588574d-d6k8n" podStartSLOduration=3.688300619 podStartE2EDuration="3.688300619s" podCreationTimestamp="2026-01-22 11:46:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:46:06.681789663 +0000 UTC m=+340.526860733" watchObservedRunningTime="2026-01-22 11:46:06.688300619 +0000 UTC m=+340.533371689" Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.934271 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g952h"] Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.936048 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.937797 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 22 11:46:06 crc kubenswrapper[4874]: I0122 11:46:06.945816 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g952h"] Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.107679 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m2vz\" (UniqueName: \"kubernetes.io/projected/1b32604b-2210-43c4-9b45-a833b8bdff64-kube-api-access-4m2vz\") pod \"community-operators-g952h\" (UID: \"1b32604b-2210-43c4-9b45-a833b8bdff64\") " pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.107759 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b32604b-2210-43c4-9b45-a833b8bdff64-utilities\") pod \"community-operators-g952h\" (UID: \"1b32604b-2210-43c4-9b45-a833b8bdff64\") " pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.107848 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b32604b-2210-43c4-9b45-a833b8bdff64-catalog-content\") pod \"community-operators-g952h\" (UID: \"1b32604b-2210-43c4-9b45-a833b8bdff64\") " pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.209152 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b32604b-2210-43c4-9b45-a833b8bdff64-catalog-content\") pod \"community-operators-g952h\" (UID: \"1b32604b-2210-43c4-9b45-a833b8bdff64\") " pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.209297 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m2vz\" (UniqueName: \"kubernetes.io/projected/1b32604b-2210-43c4-9b45-a833b8bdff64-kube-api-access-4m2vz\") pod \"community-operators-g952h\" (UID: \"1b32604b-2210-43c4-9b45-a833b8bdff64\") " pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.209447 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b32604b-2210-43c4-9b45-a833b8bdff64-utilities\") pod \"community-operators-g952h\" (UID: \"1b32604b-2210-43c4-9b45-a833b8bdff64\") " pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.210900 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b32604b-2210-43c4-9b45-a833b8bdff64-utilities\") pod \"community-operators-g952h\" (UID: \"1b32604b-2210-43c4-9b45-a833b8bdff64\") " pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.210964 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b32604b-2210-43c4-9b45-a833b8bdff64-catalog-content\") pod \"community-operators-g952h\" (UID: \"1b32604b-2210-43c4-9b45-a833b8bdff64\") " pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.226913 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m2vz\" (UniqueName: \"kubernetes.io/projected/1b32604b-2210-43c4-9b45-a833b8bdff64-kube-api-access-4m2vz\") pod \"community-operators-g952h\" (UID: \"1b32604b-2210-43c4-9b45-a833b8bdff64\") " pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.260461 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.537376 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bbdq6"] Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.538597 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.540893 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.556025 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bbdq6"] Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.614443 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlqxz\" (UniqueName: \"kubernetes.io/projected/b0a069d7-0600-4490-82a9-2656913f35b7-kube-api-access-jlqxz\") pod \"certified-operators-bbdq6\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.617485 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-utilities\") pod \"certified-operators-bbdq6\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.617875 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-catalog-content\") pod \"certified-operators-bbdq6\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.635805 4874 generic.go:334] "Generic (PLEG): container finished" podID="37153c5d-6533-4973-99bd-f682b3c148d9" containerID="c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb" exitCode=0 Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.636032 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s7gc" event={"ID":"37153c5d-6533-4973-99bd-f682b3c148d9","Type":"ContainerDied","Data":"c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb"} Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.666345 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g952h"] Jan 22 11:46:07 crc kubenswrapper[4874]: W0122 11:46:07.680162 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b32604b_2210_43c4_9b45_a833b8bdff64.slice/crio-9b914680fd3cd46302939b75768100737185d6b054310552b00b181c4e42a281 WatchSource:0}: Error finding container 9b914680fd3cd46302939b75768100737185d6b054310552b00b181c4e42a281: Status 404 returned error can't find the container with id 9b914680fd3cd46302939b75768100737185d6b054310552b00b181c4e42a281 Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.719054 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-utilities\") pod \"certified-operators-bbdq6\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.719163 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-catalog-content\") pod \"certified-operators-bbdq6\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.719200 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlqxz\" (UniqueName: \"kubernetes.io/projected/b0a069d7-0600-4490-82a9-2656913f35b7-kube-api-access-jlqxz\") pod \"certified-operators-bbdq6\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.719595 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-utilities\") pod \"certified-operators-bbdq6\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.719845 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-catalog-content\") pod \"certified-operators-bbdq6\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.737452 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlqxz\" (UniqueName: \"kubernetes.io/projected/b0a069d7-0600-4490-82a9-2656913f35b7-kube-api-access-jlqxz\") pod \"certified-operators-bbdq6\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:07 crc kubenswrapper[4874]: I0122 11:46:07.891573 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:08 crc kubenswrapper[4874]: I0122 11:46:08.273186 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bbdq6"] Jan 22 11:46:08 crc kubenswrapper[4874]: W0122 11:46:08.279874 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0a069d7_0600_4490_82a9_2656913f35b7.slice/crio-77b7f983a0caf174b2a1ff9585d546c476923723f895d8d2ccf6e1a80d72b6c1 WatchSource:0}: Error finding container 77b7f983a0caf174b2a1ff9585d546c476923723f895d8d2ccf6e1a80d72b6c1: Status 404 returned error can't find the container with id 77b7f983a0caf174b2a1ff9585d546c476923723f895d8d2ccf6e1a80d72b6c1 Jan 22 11:46:08 crc kubenswrapper[4874]: I0122 11:46:08.643574 4874 generic.go:334] "Generic (PLEG): container finished" podID="1b32604b-2210-43c4-9b45-a833b8bdff64" containerID="5cdf975ced7eb78999c4d49ef3319fe4454449901745d48f99503240c68cab9a" exitCode=0 Jan 22 11:46:08 crc kubenswrapper[4874]: I0122 11:46:08.643639 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g952h" event={"ID":"1b32604b-2210-43c4-9b45-a833b8bdff64","Type":"ContainerDied","Data":"5cdf975ced7eb78999c4d49ef3319fe4454449901745d48f99503240c68cab9a"} Jan 22 11:46:08 crc kubenswrapper[4874]: I0122 11:46:08.643878 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g952h" event={"ID":"1b32604b-2210-43c4-9b45-a833b8bdff64","Type":"ContainerStarted","Data":"9b914680fd3cd46302939b75768100737185d6b054310552b00b181c4e42a281"} Jan 22 11:46:08 crc kubenswrapper[4874]: I0122 11:46:08.646027 4874 generic.go:334] "Generic (PLEG): container finished" podID="b0a069d7-0600-4490-82a9-2656913f35b7" containerID="2136577bcdcbdcffa2ef9ace83cb7011df13fa4d06b362c2ac9b0ca6f8650b12" exitCode=0 Jan 22 11:46:08 crc kubenswrapper[4874]: I0122 11:46:08.646088 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbdq6" event={"ID":"b0a069d7-0600-4490-82a9-2656913f35b7","Type":"ContainerDied","Data":"2136577bcdcbdcffa2ef9ace83cb7011df13fa4d06b362c2ac9b0ca6f8650b12"} Jan 22 11:46:08 crc kubenswrapper[4874]: I0122 11:46:08.646106 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbdq6" event={"ID":"b0a069d7-0600-4490-82a9-2656913f35b7","Type":"ContainerStarted","Data":"77b7f983a0caf174b2a1ff9585d546c476923723f895d8d2ccf6e1a80d72b6c1"} Jan 22 11:46:08 crc kubenswrapper[4874]: I0122 11:46:08.649579 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s7gc" event={"ID":"37153c5d-6533-4973-99bd-f682b3c148d9","Type":"ContainerStarted","Data":"54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a"} Jan 22 11:46:08 crc kubenswrapper[4874]: I0122 11:46:08.653165 4874 generic.go:334] "Generic (PLEG): container finished" podID="0684b3da-a3b2-496b-be2e-b8e0fe8e1277" containerID="b9b2382b73113a44e92fe83413023a7ead8e974ad1dea0f23af30036a056af60" exitCode=0 Jan 22 11:46:08 crc kubenswrapper[4874]: I0122 11:46:08.653234 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qhz5" event={"ID":"0684b3da-a3b2-496b-be2e-b8e0fe8e1277","Type":"ContainerDied","Data":"b9b2382b73113a44e92fe83413023a7ead8e974ad1dea0f23af30036a056af60"} Jan 22 11:46:08 crc kubenswrapper[4874]: I0122 11:46:08.718168 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5s7gc" podStartSLOduration=2.214117203 podStartE2EDuration="4.718154635s" podCreationTimestamp="2026-01-22 11:46:04 +0000 UTC" firstStartedPulling="2026-01-22 11:46:05.611839725 +0000 UTC m=+339.456910795" lastFinishedPulling="2026-01-22 11:46:08.115877157 +0000 UTC m=+341.960948227" observedRunningTime="2026-01-22 11:46:08.716291386 +0000 UTC m=+342.561362496" watchObservedRunningTime="2026-01-22 11:46:08.718154635 +0000 UTC m=+342.563225705" Jan 22 11:46:09 crc kubenswrapper[4874]: I0122 11:46:09.659707 4874 generic.go:334] "Generic (PLEG): container finished" podID="b0a069d7-0600-4490-82a9-2656913f35b7" containerID="1ca6b3f9986d9290ed89e908e86e7d2e92aa24a906749df6d2ac85d44a6ba2a1" exitCode=0 Jan 22 11:46:09 crc kubenswrapper[4874]: I0122 11:46:09.659777 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbdq6" event={"ID":"b0a069d7-0600-4490-82a9-2656913f35b7","Type":"ContainerDied","Data":"1ca6b3f9986d9290ed89e908e86e7d2e92aa24a906749df6d2ac85d44a6ba2a1"} Jan 22 11:46:09 crc kubenswrapper[4874]: I0122 11:46:09.671421 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qhz5" event={"ID":"0684b3da-a3b2-496b-be2e-b8e0fe8e1277","Type":"ContainerStarted","Data":"57f11e14580c8438fa69024b2c6e29e0e1d4e35056bf6b184f9ec8980bae6bb9"} Jan 22 11:46:09 crc kubenswrapper[4874]: I0122 11:46:09.674080 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g952h" event={"ID":"1b32604b-2210-43c4-9b45-a833b8bdff64","Type":"ContainerStarted","Data":"5b17401120283f26aec712f87e671ae70042045fd1b29eda8aa624834e6d83ad"} Jan 22 11:46:09 crc kubenswrapper[4874]: I0122 11:46:09.733917 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7qhz5" podStartSLOduration=2.311162537 podStartE2EDuration="4.733899578s" podCreationTimestamp="2026-01-22 11:46:05 +0000 UTC" firstStartedPulling="2026-01-22 11:46:06.625999599 +0000 UTC m=+340.471070669" lastFinishedPulling="2026-01-22 11:46:09.04873664 +0000 UTC m=+342.893807710" observedRunningTime="2026-01-22 11:46:09.728702324 +0000 UTC m=+343.573773414" watchObservedRunningTime="2026-01-22 11:46:09.733899578 +0000 UTC m=+343.578970668" Jan 22 11:46:09 crc kubenswrapper[4874]: I0122 11:46:09.832476 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-n2lnd" Jan 22 11:46:09 crc kubenswrapper[4874]: I0122 11:46:09.940457 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jg4wj"] Jan 22 11:46:10 crc kubenswrapper[4874]: I0122 11:46:10.679354 4874 generic.go:334] "Generic (PLEG): container finished" podID="1b32604b-2210-43c4-9b45-a833b8bdff64" containerID="5b17401120283f26aec712f87e671ae70042045fd1b29eda8aa624834e6d83ad" exitCode=0 Jan 22 11:46:10 crc kubenswrapper[4874]: I0122 11:46:10.679453 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g952h" event={"ID":"1b32604b-2210-43c4-9b45-a833b8bdff64","Type":"ContainerDied","Data":"5b17401120283f26aec712f87e671ae70042045fd1b29eda8aa624834e6d83ad"} Jan 22 11:46:10 crc kubenswrapper[4874]: I0122 11:46:10.682293 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbdq6" event={"ID":"b0a069d7-0600-4490-82a9-2656913f35b7","Type":"ContainerStarted","Data":"56ec71b649678b9d06d64da2f108ecc5d7baa2ab0550b12655288c4bbfbcdef7"} Jan 22 11:46:10 crc kubenswrapper[4874]: I0122 11:46:10.716735 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bbdq6" podStartSLOduration=2.244238633 podStartE2EDuration="3.716716741s" podCreationTimestamp="2026-01-22 11:46:07 +0000 UTC" firstStartedPulling="2026-01-22 11:46:08.64748192 +0000 UTC m=+342.492552990" lastFinishedPulling="2026-01-22 11:46:10.119960028 +0000 UTC m=+343.965031098" observedRunningTime="2026-01-22 11:46:10.714410068 +0000 UTC m=+344.559481148" watchObservedRunningTime="2026-01-22 11:46:10.716716741 +0000 UTC m=+344.561787811" Jan 22 11:46:11 crc kubenswrapper[4874]: I0122 11:46:11.691528 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g952h" event={"ID":"1b32604b-2210-43c4-9b45-a833b8bdff64","Type":"ContainerStarted","Data":"fc909630936b38a104b45954668e2ed68928a36f5e77acab9c59012caae6c2e0"} Jan 22 11:46:13 crc kubenswrapper[4874]: I0122 11:46:13.520949 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:46:13 crc kubenswrapper[4874]: I0122 11:46:13.521011 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:46:14 crc kubenswrapper[4874]: I0122 11:46:14.861387 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:14 crc kubenswrapper[4874]: I0122 11:46:14.861753 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:14 crc kubenswrapper[4874]: I0122 11:46:14.912180 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:14 crc kubenswrapper[4874]: I0122 11:46:14.928347 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g952h" podStartSLOduration=6.395079881 podStartE2EDuration="8.928325416s" podCreationTimestamp="2026-01-22 11:46:06 +0000 UTC" firstStartedPulling="2026-01-22 11:46:08.645137936 +0000 UTC m=+342.490209006" lastFinishedPulling="2026-01-22 11:46:11.178383471 +0000 UTC m=+345.023454541" observedRunningTime="2026-01-22 11:46:11.709976313 +0000 UTC m=+345.555047383" watchObservedRunningTime="2026-01-22 11:46:14.928325416 +0000 UTC m=+348.773396486" Jan 22 11:46:15 crc kubenswrapper[4874]: I0122 11:46:15.481095 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:15 crc kubenswrapper[4874]: I0122 11:46:15.481495 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:15 crc kubenswrapper[4874]: I0122 11:46:15.516481 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:15 crc kubenswrapper[4874]: I0122 11:46:15.749054 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:46:15 crc kubenswrapper[4874]: I0122 11:46:15.758638 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7qhz5" Jan 22 11:46:17 crc kubenswrapper[4874]: I0122 11:46:17.261096 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:17 crc kubenswrapper[4874]: I0122 11:46:17.261179 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:17 crc kubenswrapper[4874]: I0122 11:46:17.332391 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:17 crc kubenswrapper[4874]: I0122 11:46:17.781105 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g952h" Jan 22 11:46:17 crc kubenswrapper[4874]: I0122 11:46:17.892116 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:17 crc kubenswrapper[4874]: I0122 11:46:17.892179 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:17 crc kubenswrapper[4874]: I0122 11:46:17.941246 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:18 crc kubenswrapper[4874]: I0122 11:46:18.804489 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:46:34 crc kubenswrapper[4874]: I0122 11:46:34.980012 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" podUID="90a037cd-cacf-4706-a857-f65c8f16c384" containerName="registry" containerID="cri-o://e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe" gracePeriod=30 Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.420557 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.506265 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/90a037cd-cacf-4706-a857-f65c8f16c384-ca-trust-extracted\") pod \"90a037cd-cacf-4706-a857-f65c8f16c384\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.506418 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/90a037cd-cacf-4706-a857-f65c8f16c384-installation-pull-secrets\") pod \"90a037cd-cacf-4706-a857-f65c8f16c384\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.506464 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-registry-tls\") pod \"90a037cd-cacf-4706-a857-f65c8f16c384\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.506504 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-bound-sa-token\") pod \"90a037cd-cacf-4706-a857-f65c8f16c384\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.506526 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-registry-certificates\") pod \"90a037cd-cacf-4706-a857-f65c8f16c384\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.506541 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwtlg\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-kube-api-access-zwtlg\") pod \"90a037cd-cacf-4706-a857-f65c8f16c384\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.506717 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"90a037cd-cacf-4706-a857-f65c8f16c384\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.506745 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-trusted-ca\") pod \"90a037cd-cacf-4706-a857-f65c8f16c384\" (UID: \"90a037cd-cacf-4706-a857-f65c8f16c384\") " Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.507609 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "90a037cd-cacf-4706-a857-f65c8f16c384" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.508604 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "90a037cd-cacf-4706-a857-f65c8f16c384" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.513697 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "90a037cd-cacf-4706-a857-f65c8f16c384" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.514086 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-kube-api-access-zwtlg" (OuterVolumeSpecName: "kube-api-access-zwtlg") pod "90a037cd-cacf-4706-a857-f65c8f16c384" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384"). InnerVolumeSpecName "kube-api-access-zwtlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.514703 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90a037cd-cacf-4706-a857-f65c8f16c384-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "90a037cd-cacf-4706-a857-f65c8f16c384" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.515377 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "90a037cd-cacf-4706-a857-f65c8f16c384" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.526300 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "90a037cd-cacf-4706-a857-f65c8f16c384" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.535469 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90a037cd-cacf-4706-a857-f65c8f16c384-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "90a037cd-cacf-4706-a857-f65c8f16c384" (UID: "90a037cd-cacf-4706-a857-f65c8f16c384"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.608670 4874 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.608704 4874 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.608715 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwtlg\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-kube-api-access-zwtlg\") on node \"crc\" DevicePath \"\"" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.608724 4874 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90a037cd-cacf-4706-a857-f65c8f16c384-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.608733 4874 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/90a037cd-cacf-4706-a857-f65c8f16c384-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.608743 4874 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/90a037cd-cacf-4706-a857-f65c8f16c384-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.608752 4874 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/90a037cd-cacf-4706-a857-f65c8f16c384-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.828302 4874 generic.go:334] "Generic (PLEG): container finished" podID="90a037cd-cacf-4706-a857-f65c8f16c384" containerID="e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe" exitCode=0 Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.828381 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.828378 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" event={"ID":"90a037cd-cacf-4706-a857-f65c8f16c384","Type":"ContainerDied","Data":"e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe"} Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.828475 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jg4wj" event={"ID":"90a037cd-cacf-4706-a857-f65c8f16c384","Type":"ContainerDied","Data":"61abb1ca8d94b8a93a2470a92da41ddacd13e2ac8b6f3b50b461fb0f48f44b61"} Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.828495 4874 scope.go:117] "RemoveContainer" containerID="e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.852822 4874 scope.go:117] "RemoveContainer" containerID="e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe" Jan 22 11:46:35 crc kubenswrapper[4874]: E0122 11:46:35.853340 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe\": container with ID starting with e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe not found: ID does not exist" containerID="e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.853370 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe"} err="failed to get container status \"e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe\": rpc error: code = NotFound desc = could not find container \"e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe\": container with ID starting with e38ad0098b8ea8cfa9bacebe690cc8d3c60a195c4c2554e6c860332c9db56abe not found: ID does not exist" Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.890493 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jg4wj"] Jan 22 11:46:35 crc kubenswrapper[4874]: I0122 11:46:35.898148 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jg4wj"] Jan 22 11:46:36 crc kubenswrapper[4874]: I0122 11:46:36.722931 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90a037cd-cacf-4706-a857-f65c8f16c384" path="/var/lib/kubelet/pods/90a037cd-cacf-4706-a857-f65c8f16c384/volumes" Jan 22 11:46:43 crc kubenswrapper[4874]: I0122 11:46:43.520095 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:46:43 crc kubenswrapper[4874]: I0122 11:46:43.520849 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:47:13 crc kubenswrapper[4874]: I0122 11:47:13.520332 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:47:13 crc kubenswrapper[4874]: I0122 11:47:13.522802 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:47:13 crc kubenswrapper[4874]: I0122 11:47:13.523038 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:47:13 crc kubenswrapper[4874]: I0122 11:47:13.524066 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf675adc04e058930041f77cfb016f23f15475800ec1dca3cd6db1579e71257b"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 11:47:13 crc kubenswrapper[4874]: I0122 11:47:13.524523 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://cf675adc04e058930041f77cfb016f23f15475800ec1dca3cd6db1579e71257b" gracePeriod=600 Jan 22 11:47:14 crc kubenswrapper[4874]: I0122 11:47:14.087648 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="cf675adc04e058930041f77cfb016f23f15475800ec1dca3cd6db1579e71257b" exitCode=0 Jan 22 11:47:14 crc kubenswrapper[4874]: I0122 11:47:14.088009 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"cf675adc04e058930041f77cfb016f23f15475800ec1dca3cd6db1579e71257b"} Jan 22 11:47:14 crc kubenswrapper[4874]: I0122 11:47:14.088040 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"4ef8bf2a0eb4af528f4de7cda955aa92b0de433bdf6d0ebda36103e1834fe2b2"} Jan 22 11:47:14 crc kubenswrapper[4874]: I0122 11:47:14.088059 4874 scope.go:117] "RemoveContainer" containerID="e71ba793b9f7a76acd2521c1621a1ae38890f23c9caaa3d033e416369701b192" Jan 22 11:49:13 crc kubenswrapper[4874]: I0122 11:49:13.520259 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:49:13 crc kubenswrapper[4874]: I0122 11:49:13.520747 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:49:43 crc kubenswrapper[4874]: I0122 11:49:43.520507 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:49:43 crc kubenswrapper[4874]: I0122 11:49:43.520992 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:50:13 crc kubenswrapper[4874]: I0122 11:50:13.520779 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:50:13 crc kubenswrapper[4874]: I0122 11:50:13.522607 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:50:13 crc kubenswrapper[4874]: I0122 11:50:13.522693 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:50:13 crc kubenswrapper[4874]: I0122 11:50:13.523345 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ef8bf2a0eb4af528f4de7cda955aa92b0de433bdf6d0ebda36103e1834fe2b2"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 11:50:13 crc kubenswrapper[4874]: I0122 11:50:13.523437 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://4ef8bf2a0eb4af528f4de7cda955aa92b0de433bdf6d0ebda36103e1834fe2b2" gracePeriod=600 Jan 22 11:50:14 crc kubenswrapper[4874]: I0122 11:50:14.282086 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="4ef8bf2a0eb4af528f4de7cda955aa92b0de433bdf6d0ebda36103e1834fe2b2" exitCode=0 Jan 22 11:50:14 crc kubenswrapper[4874]: I0122 11:50:14.282141 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"4ef8bf2a0eb4af528f4de7cda955aa92b0de433bdf6d0ebda36103e1834fe2b2"} Jan 22 11:50:14 crc kubenswrapper[4874]: I0122 11:50:14.282691 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"58a69e8f9170bdd4dd90e6e773cd03089d2a6279398d2b1a2ba4ed87135be13a"} Jan 22 11:50:14 crc kubenswrapper[4874]: I0122 11:50:14.282719 4874 scope.go:117] "RemoveContainer" containerID="cf675adc04e058930041f77cfb016f23f15475800ec1dca3cd6db1579e71257b" Jan 22 11:51:05 crc kubenswrapper[4874]: I0122 11:51:05.774178 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6tmll"] Jan 22 11:51:05 crc kubenswrapper[4874]: I0122 11:51:05.775083 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovn-controller" containerID="cri-o://3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794" gracePeriod=30 Jan 22 11:51:05 crc kubenswrapper[4874]: I0122 11:51:05.775499 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="sbdb" containerID="cri-o://c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de" gracePeriod=30 Jan 22 11:51:05 crc kubenswrapper[4874]: I0122 11:51:05.775539 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="nbdb" containerID="cri-o://eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079" gracePeriod=30 Jan 22 11:51:05 crc kubenswrapper[4874]: I0122 11:51:05.775572 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="northd" containerID="cri-o://6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6" gracePeriod=30 Jan 22 11:51:05 crc kubenswrapper[4874]: I0122 11:51:05.775598 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2" gracePeriod=30 Jan 22 11:51:05 crc kubenswrapper[4874]: I0122 11:51:05.775624 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="kube-rbac-proxy-node" containerID="cri-o://32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348" gracePeriod=30 Jan 22 11:51:05 crc kubenswrapper[4874]: I0122 11:51:05.775654 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovn-acl-logging" containerID="cri-o://af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba" gracePeriod=30 Jan 22 11:51:05 crc kubenswrapper[4874]: I0122 11:51:05.825235 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" containerID="cri-o://a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161" gracePeriod=30 Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.113356 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/3.log" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.116304 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovn-acl-logging/0.log" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.116895 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovn-controller/0.log" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.117320 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.185948 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lbtk2"] Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186225 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovn-acl-logging" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186257 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovn-acl-logging" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186277 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="kube-rbac-proxy-node" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186289 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="kube-rbac-proxy-node" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186306 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="sbdb" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186316 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="sbdb" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186330 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186340 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186353 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="kube-rbac-proxy-ovn-metrics" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186365 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="kube-rbac-proxy-ovn-metrics" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186381 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovn-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186392 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovn-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186437 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186449 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186463 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186472 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186485 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="kubecfg-setup" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186495 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="kubecfg-setup" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186515 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="nbdb" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186525 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="nbdb" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186542 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="northd" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186553 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="northd" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.186585 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90a037cd-cacf-4706-a857-f65c8f16c384" containerName="registry" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186596 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="90a037cd-cacf-4706-a857-f65c8f16c384" containerName="registry" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186770 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovn-acl-logging" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186784 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186796 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186806 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="nbdb" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186823 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="northd" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186839 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="90a037cd-cacf-4706-a857-f65c8f16c384" containerName="registry" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186854 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="sbdb" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186868 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="kube-rbac-proxy-ovn-metrics" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186882 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186894 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovn-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.186906 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="kube-rbac-proxy-node" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.187065 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.187081 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.187264 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.187449 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.187468 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.187605 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerName="ovnkube-controller" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.189689 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.277778 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9lv4\" (UniqueName: \"kubernetes.io/projected/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-kube-api-access-t9lv4\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.277849 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-netns\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.277922 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-ovn\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.277960 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-openvswitch\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.277995 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-netd\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278035 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-bin\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278066 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-systemd-units\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278100 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-var-lib-openvswitch\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278133 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-etc-openvswitch\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278126 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278155 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278177 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-env-overrides\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278232 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278266 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-config\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278248 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278281 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278292 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278301 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovn-node-metrics-cert\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278274 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278350 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278508 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-script-lib\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278582 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-node-log\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278668 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-ovn-kubernetes\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278692 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-node-log" (OuterVolumeSpecName: "node-log") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278694 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278727 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278739 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278781 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-log-socket\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278791 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278789 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278861 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-log-socket" (OuterVolumeSpecName: "log-socket") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278825 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-systemd\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278924 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-kubelet\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.278965 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-slash\") pod \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\" (UID: \"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9\") " Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279010 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279119 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-slash" (OuterVolumeSpecName: "host-slash") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279242 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-run-openvswitch\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279313 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-run-systemd\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279313 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279352 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-cni-bin\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279384 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-node-log\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279476 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-slash\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279521 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5937262c-c601-4bdc-9e86-06900a448320-ovnkube-config\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279554 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5937262c-c601-4bdc-9e86-06900a448320-ovn-node-metrics-cert\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279586 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-run-ovn-kubernetes\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279636 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5937262c-c601-4bdc-9e86-06900a448320-env-overrides\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279671 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5937262c-c601-4bdc-9e86-06900a448320-ovnkube-script-lib\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279712 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-var-lib-openvswitch\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279749 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-run-ovn\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279823 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-cni-netd\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279857 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279899 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-kubelet\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.279935 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2vl7\" (UniqueName: \"kubernetes.io/projected/5937262c-c601-4bdc-9e86-06900a448320-kube-api-access-n2vl7\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280050 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-log-socket\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280103 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-etc-openvswitch\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280193 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-systemd-units\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280251 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-run-netns\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280345 4874 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280369 4874 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280390 4874 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280438 4874 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280458 4874 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280476 4874 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280494 4874 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280512 4874 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280528 4874 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280545 4874 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280566 4874 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-node-log\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280583 4874 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280603 4874 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280620 4874 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-log-socket\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280638 4874 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280654 4874 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-slash\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.280671 4874 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.284911 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-kube-api-access-t9lv4" (OuterVolumeSpecName: "kube-api-access-t9lv4") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "kube-api-access-t9lv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.285822 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.294164 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" (UID: "642d0ca0-2e0f-4b69-9484-a63d0a01f8a9"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.381931 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-run-openvswitch\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382041 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-run-systemd\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382086 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-cni-bin\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382133 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-node-log\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382189 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-slash\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382240 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5937262c-c601-4bdc-9e86-06900a448320-ovnkube-config\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382283 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5937262c-c601-4bdc-9e86-06900a448320-ovn-node-metrics-cert\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382331 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-run-ovn-kubernetes\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382374 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5937262c-c601-4bdc-9e86-06900a448320-env-overrides\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382461 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5937262c-c601-4bdc-9e86-06900a448320-ovnkube-script-lib\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382514 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-var-lib-openvswitch\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382561 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-run-ovn\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382638 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-cni-netd\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382721 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382782 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-kubelet\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382824 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2vl7\" (UniqueName: \"kubernetes.io/projected/5937262c-c601-4bdc-9e86-06900a448320-kube-api-access-n2vl7\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382865 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-log-socket\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382908 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-etc-openvswitch\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.382972 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-systemd-units\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383025 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-run-netns\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383135 4874 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383174 4874 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383202 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9lv4\" (UniqueName: \"kubernetes.io/projected/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9-kube-api-access-t9lv4\") on node \"crc\" DevicePath \"\"" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383290 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-run-netns\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383373 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-run-openvswitch\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383472 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-run-systemd\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383530 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-cni-bin\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383589 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-run-ovn\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383682 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-var-lib-openvswitch\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383704 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-etc-openvswitch\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383754 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-log-socket\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383827 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-run-ovn-kubernetes\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383872 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-kubelet\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383827 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383949 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-slash\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383997 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-node-log\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.383874 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-host-cni-netd\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.384200 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5937262c-c601-4bdc-9e86-06900a448320-systemd-units\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.385003 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5937262c-c601-4bdc-9e86-06900a448320-ovnkube-script-lib\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.385107 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5937262c-c601-4bdc-9e86-06900a448320-env-overrides\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.385592 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5937262c-c601-4bdc-9e86-06900a448320-ovnkube-config\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.390789 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5937262c-c601-4bdc-9e86-06900a448320-ovn-node-metrics-cert\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.413653 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2vl7\" (UniqueName: \"kubernetes.io/projected/5937262c-c601-4bdc-9e86-06900a448320-kube-api-access-n2vl7\") pod \"ovnkube-node-lbtk2\" (UID: \"5937262c-c601-4bdc-9e86-06900a448320\") " pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.509490 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:06 crc kubenswrapper[4874]: W0122 11:51:06.544141 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5937262c_c601_4bdc_9e86_06900a448320.slice/crio-fbc8c59e1ba96d09e2275be4b4fef1829c84d16c3b415161dad6e72827b47264 WatchSource:0}: Error finding container fbc8c59e1ba96d09e2275be4b4fef1829c84d16c3b415161dad6e72827b47264: Status 404 returned error can't find the container with id fbc8c59e1ba96d09e2275be4b4fef1829c84d16c3b415161dad6e72827b47264 Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.656811 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" event={"ID":"5937262c-c601-4bdc-9e86-06900a448320","Type":"ContainerStarted","Data":"fbc8c59e1ba96d09e2275be4b4fef1829c84d16c3b415161dad6e72827b47264"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.662355 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovnkube-controller/3.log" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.669034 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovn-acl-logging/0.log" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.669760 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6tmll_642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/ovn-controller/0.log" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670356 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161" exitCode=0 Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670436 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de" exitCode=0 Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670457 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079" exitCode=0 Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670472 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6" exitCode=0 Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670487 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2" exitCode=0 Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670470 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670554 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670580 4874 scope.go:117] "RemoveContainer" containerID="a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670503 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348" exitCode=0 Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670653 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba" exitCode=143 Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670669 4874 generic.go:334] "Generic (PLEG): container finished" podID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" containerID="3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794" exitCode=143 Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670559 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670742 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670770 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670788 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670804 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670845 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670866 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670900 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670907 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670933 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670978 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670985 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670992 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.670998 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671008 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671020 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671028 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671034 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671040 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671046 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671052 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671060 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671066 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671155 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671165 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671175 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671187 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671195 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671201 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671207 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671214 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671506 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671524 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671531 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671537 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671544 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671555 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6tmll" event={"ID":"642d0ca0-2e0f-4b69-9484-a63d0a01f8a9","Type":"ContainerDied","Data":"1a325212afdff76674e62ed80b3cf828c221bb977e244657ca644f4020d804af"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671571 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671580 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671587 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671593 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671600 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671607 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671613 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671619 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671626 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.671632 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.673027 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/2.log" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.675806 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/1.log" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.675877 4874 generic.go:334] "Generic (PLEG): container finished" podID="977746b5-ac1b-4b6e-bdbc-ddd90225e68c" containerID="55eeb9abd8c425711e374c107c22ec24d1741880327f226d7db5e06d67925630" exitCode=2 Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.675921 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-krrtc" event={"ID":"977746b5-ac1b-4b6e-bdbc-ddd90225e68c","Type":"ContainerDied","Data":"55eeb9abd8c425711e374c107c22ec24d1741880327f226d7db5e06d67925630"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.676068 4874 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8"} Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.676487 4874 scope.go:117] "RemoveContainer" containerID="55eeb9abd8c425711e374c107c22ec24d1741880327f226d7db5e06d67925630" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.676726 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-krrtc_openshift-multus(977746b5-ac1b-4b6e-bdbc-ddd90225e68c)\"" pod="openshift-multus/multus-krrtc" podUID="977746b5-ac1b-4b6e-bdbc-ddd90225e68c" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.697296 4874 scope.go:117] "RemoveContainer" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.778073 4874 scope.go:117] "RemoveContainer" containerID="c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.785469 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6tmll"] Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.792178 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6tmll"] Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.794002 4874 scope.go:117] "RemoveContainer" containerID="eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.806172 4874 scope.go:117] "RemoveContainer" containerID="6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.821312 4874 scope.go:117] "RemoveContainer" containerID="284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.840933 4874 scope.go:117] "RemoveContainer" containerID="32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.854324 4874 scope.go:117] "RemoveContainer" containerID="af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.868230 4874 scope.go:117] "RemoveContainer" containerID="3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.882334 4874 scope.go:117] "RemoveContainer" containerID="f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.895665 4874 scope.go:117] "RemoveContainer" containerID="a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.896088 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161\": container with ID starting with a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161 not found: ID does not exist" containerID="a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.896135 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161"} err="failed to get container status \"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161\": rpc error: code = NotFound desc = could not find container \"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161\": container with ID starting with a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.896168 4874 scope.go:117] "RemoveContainer" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.896568 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\": container with ID starting with 052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35 not found: ID does not exist" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.896596 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35"} err="failed to get container status \"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\": rpc error: code = NotFound desc = could not find container \"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\": container with ID starting with 052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.896621 4874 scope.go:117] "RemoveContainer" containerID="c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.897024 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\": container with ID starting with c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de not found: ID does not exist" containerID="c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.897060 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de"} err="failed to get container status \"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\": rpc error: code = NotFound desc = could not find container \"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\": container with ID starting with c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.897074 4874 scope.go:117] "RemoveContainer" containerID="eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.897473 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\": container with ID starting with eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079 not found: ID does not exist" containerID="eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.897535 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079"} err="failed to get container status \"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\": rpc error: code = NotFound desc = could not find container \"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\": container with ID starting with eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.897571 4874 scope.go:117] "RemoveContainer" containerID="6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.898077 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\": container with ID starting with 6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6 not found: ID does not exist" containerID="6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.898122 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6"} err="failed to get container status \"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\": rpc error: code = NotFound desc = could not find container \"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\": container with ID starting with 6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.898141 4874 scope.go:117] "RemoveContainer" containerID="284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.898554 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\": container with ID starting with 284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2 not found: ID does not exist" containerID="284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.898603 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2"} err="failed to get container status \"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\": rpc error: code = NotFound desc = could not find container \"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\": container with ID starting with 284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.898633 4874 scope.go:117] "RemoveContainer" containerID="32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.898974 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\": container with ID starting with 32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348 not found: ID does not exist" containerID="32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.899013 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348"} err="failed to get container status \"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\": rpc error: code = NotFound desc = could not find container \"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\": container with ID starting with 32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.899030 4874 scope.go:117] "RemoveContainer" containerID="af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.899323 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\": container with ID starting with af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba not found: ID does not exist" containerID="af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.899354 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba"} err="failed to get container status \"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\": rpc error: code = NotFound desc = could not find container \"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\": container with ID starting with af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.899370 4874 scope.go:117] "RemoveContainer" containerID="3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.899655 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\": container with ID starting with 3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794 not found: ID does not exist" containerID="3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.899679 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794"} err="failed to get container status \"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\": rpc error: code = NotFound desc = could not find container \"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\": container with ID starting with 3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.899694 4874 scope.go:117] "RemoveContainer" containerID="f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f" Jan 22 11:51:06 crc kubenswrapper[4874]: E0122 11:51:06.899931 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\": container with ID starting with f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f not found: ID does not exist" containerID="f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.899976 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f"} err="failed to get container status \"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\": rpc error: code = NotFound desc = could not find container \"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\": container with ID starting with f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.900001 4874 scope.go:117] "RemoveContainer" containerID="a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.900276 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161"} err="failed to get container status \"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161\": rpc error: code = NotFound desc = could not find container \"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161\": container with ID starting with a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.900299 4874 scope.go:117] "RemoveContainer" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.900535 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35"} err="failed to get container status \"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\": rpc error: code = NotFound desc = could not find container \"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\": container with ID starting with 052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.900577 4874 scope.go:117] "RemoveContainer" containerID="c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.901016 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de"} err="failed to get container status \"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\": rpc error: code = NotFound desc = could not find container \"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\": container with ID starting with c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.901034 4874 scope.go:117] "RemoveContainer" containerID="eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.901361 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079"} err="failed to get container status \"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\": rpc error: code = NotFound desc = could not find container \"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\": container with ID starting with eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.901378 4874 scope.go:117] "RemoveContainer" containerID="6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.901707 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6"} err="failed to get container status \"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\": rpc error: code = NotFound desc = could not find container \"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\": container with ID starting with 6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.901740 4874 scope.go:117] "RemoveContainer" containerID="284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.902330 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2"} err="failed to get container status \"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\": rpc error: code = NotFound desc = could not find container \"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\": container with ID starting with 284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.902443 4874 scope.go:117] "RemoveContainer" containerID="32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.902982 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348"} err="failed to get container status \"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\": rpc error: code = NotFound desc = could not find container \"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\": container with ID starting with 32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.903022 4874 scope.go:117] "RemoveContainer" containerID="af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.903275 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba"} err="failed to get container status \"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\": rpc error: code = NotFound desc = could not find container \"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\": container with ID starting with af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.903297 4874 scope.go:117] "RemoveContainer" containerID="3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.903557 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794"} err="failed to get container status \"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\": rpc error: code = NotFound desc = could not find container \"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\": container with ID starting with 3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.903588 4874 scope.go:117] "RemoveContainer" containerID="f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.903849 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f"} err="failed to get container status \"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\": rpc error: code = NotFound desc = could not find container \"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\": container with ID starting with f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.903867 4874 scope.go:117] "RemoveContainer" containerID="a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.904076 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161"} err="failed to get container status \"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161\": rpc error: code = NotFound desc = could not find container \"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161\": container with ID starting with a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.904109 4874 scope.go:117] "RemoveContainer" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.904366 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35"} err="failed to get container status \"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\": rpc error: code = NotFound desc = could not find container \"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\": container with ID starting with 052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.904386 4874 scope.go:117] "RemoveContainer" containerID="c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.904645 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de"} err="failed to get container status \"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\": rpc error: code = NotFound desc = could not find container \"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\": container with ID starting with c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.904676 4874 scope.go:117] "RemoveContainer" containerID="eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.904884 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079"} err="failed to get container status \"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\": rpc error: code = NotFound desc = could not find container \"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\": container with ID starting with eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.904901 4874 scope.go:117] "RemoveContainer" containerID="6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.905187 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6"} err="failed to get container status \"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\": rpc error: code = NotFound desc = could not find container \"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\": container with ID starting with 6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.905219 4874 scope.go:117] "RemoveContainer" containerID="284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.905681 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2"} err="failed to get container status \"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\": rpc error: code = NotFound desc = could not find container \"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\": container with ID starting with 284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.905714 4874 scope.go:117] "RemoveContainer" containerID="32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.905963 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348"} err="failed to get container status \"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\": rpc error: code = NotFound desc = could not find container \"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\": container with ID starting with 32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.905982 4874 scope.go:117] "RemoveContainer" containerID="af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.906183 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba"} err="failed to get container status \"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\": rpc error: code = NotFound desc = could not find container \"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\": container with ID starting with af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.906212 4874 scope.go:117] "RemoveContainer" containerID="3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.906492 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794"} err="failed to get container status \"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\": rpc error: code = NotFound desc = could not find container \"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\": container with ID starting with 3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.906525 4874 scope.go:117] "RemoveContainer" containerID="f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.906801 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f"} err="failed to get container status \"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\": rpc error: code = NotFound desc = could not find container \"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\": container with ID starting with f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.906824 4874 scope.go:117] "RemoveContainer" containerID="a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.907138 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161"} err="failed to get container status \"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161\": rpc error: code = NotFound desc = could not find container \"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161\": container with ID starting with a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.907182 4874 scope.go:117] "RemoveContainer" containerID="052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.907504 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35"} err="failed to get container status \"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\": rpc error: code = NotFound desc = could not find container \"052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35\": container with ID starting with 052dc6d3925ff9785ba5a4c833cc761b29f083abcaeff3104c5c103c65659b35 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.907524 4874 scope.go:117] "RemoveContainer" containerID="c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.907838 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de"} err="failed to get container status \"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\": rpc error: code = NotFound desc = could not find container \"c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de\": container with ID starting with c8820ecf159f799b4b40af886c3788127aacb2e9f59ae4e5f53af1798da898de not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.907886 4874 scope.go:117] "RemoveContainer" containerID="eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.908337 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079"} err="failed to get container status \"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\": rpc error: code = NotFound desc = could not find container \"eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079\": container with ID starting with eabf4bd0bcaadf2006582497f2d0e6aaf0ff5b4d1be038a58e45e3484e629079 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.908355 4874 scope.go:117] "RemoveContainer" containerID="6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.908618 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6"} err="failed to get container status \"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\": rpc error: code = NotFound desc = could not find container \"6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6\": container with ID starting with 6b440f662ef1c69071309de21c1a20246fd2ae4d01c92086287391a6fe5848d6 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.908656 4874 scope.go:117] "RemoveContainer" containerID="284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.909015 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2"} err="failed to get container status \"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\": rpc error: code = NotFound desc = could not find container \"284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2\": container with ID starting with 284368ec533c5ce597410ad825babc3c53211edce42a7298b0b78d38c510b6f2 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.909050 4874 scope.go:117] "RemoveContainer" containerID="32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.909595 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348"} err="failed to get container status \"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\": rpc error: code = NotFound desc = could not find container \"32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348\": container with ID starting with 32a81acf1a16e21ef8fa656065235d38beb28ea4d4dc28f8519d1d9f10347348 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.909640 4874 scope.go:117] "RemoveContainer" containerID="af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.910179 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba"} err="failed to get container status \"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\": rpc error: code = NotFound desc = could not find container \"af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba\": container with ID starting with af3ca476362022cf4305d74c7f29819abc8a48d52222d171a3056f10e2cb67ba not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.910216 4874 scope.go:117] "RemoveContainer" containerID="3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.910577 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794"} err="failed to get container status \"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\": rpc error: code = NotFound desc = could not find container \"3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794\": container with ID starting with 3d0bdbbba76909afd7862ef9e9ef224eec57d72779dab9160af584b279e14794 not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.910607 4874 scope.go:117] "RemoveContainer" containerID="f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.910927 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f"} err="failed to get container status \"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\": rpc error: code = NotFound desc = could not find container \"f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f\": container with ID starting with f849b97108e8ca007253cf2f4a548c6362f9618e9110a0d4b19e99c0ade6de1f not found: ID does not exist" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.911006 4874 scope.go:117] "RemoveContainer" containerID="a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161" Jan 22 11:51:06 crc kubenswrapper[4874]: I0122 11:51:06.911334 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161"} err="failed to get container status \"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161\": rpc error: code = NotFound desc = could not find container \"a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161\": container with ID starting with a6b6cc9e9262eb8c84117af65760c4bb3d7625a0b7b8b0c3c3eefbcde0b59161 not found: ID does not exist" Jan 22 11:51:07 crc kubenswrapper[4874]: I0122 11:51:07.688237 4874 generic.go:334] "Generic (PLEG): container finished" podID="5937262c-c601-4bdc-9e86-06900a448320" containerID="3ac568eb1b73e9db0db4c813128a52165babb0168916650443a514d5161194ef" exitCode=0 Jan 22 11:51:07 crc kubenswrapper[4874]: I0122 11:51:07.688633 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" event={"ID":"5937262c-c601-4bdc-9e86-06900a448320","Type":"ContainerDied","Data":"3ac568eb1b73e9db0db4c813128a52165babb0168916650443a514d5161194ef"} Jan 22 11:51:08 crc kubenswrapper[4874]: I0122 11:51:08.701148 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" event={"ID":"5937262c-c601-4bdc-9e86-06900a448320","Type":"ContainerStarted","Data":"734d12487de38e83a56ed62534d303f239155ed627c43e63605ed1673cd0b059"} Jan 22 11:51:08 crc kubenswrapper[4874]: I0122 11:51:08.701680 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" event={"ID":"5937262c-c601-4bdc-9e86-06900a448320","Type":"ContainerStarted","Data":"cbb55d3d4711f04a0f6114d8083191131c8f713b613cab320ced8fe2bb210ff7"} Jan 22 11:51:08 crc kubenswrapper[4874]: I0122 11:51:08.701718 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" event={"ID":"5937262c-c601-4bdc-9e86-06900a448320","Type":"ContainerStarted","Data":"612190b71092305c4861a3b9c5639a0b0a1d425f2488415b6aecfae090a5535b"} Jan 22 11:51:08 crc kubenswrapper[4874]: I0122 11:51:08.701745 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" event={"ID":"5937262c-c601-4bdc-9e86-06900a448320","Type":"ContainerStarted","Data":"dbf962f2aeb6324d88ba5f19b6dd0a7037260c1fda9bba552aa5be254db0246d"} Jan 22 11:51:08 crc kubenswrapper[4874]: I0122 11:51:08.701770 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" event={"ID":"5937262c-c601-4bdc-9e86-06900a448320","Type":"ContainerStarted","Data":"84ef83580edfe2cf2b0d6af52e89ea6977f864611b4366deb801a90212c3a208"} Jan 22 11:51:08 crc kubenswrapper[4874]: I0122 11:51:08.701795 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" event={"ID":"5937262c-c601-4bdc-9e86-06900a448320","Type":"ContainerStarted","Data":"1f0e69d384556682f0a74cbe00ddd73f9942f812445322bb19395eba5cd1705a"} Jan 22 11:51:08 crc kubenswrapper[4874]: I0122 11:51:08.730602 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="642d0ca0-2e0f-4b69-9484-a63d0a01f8a9" path="/var/lib/kubelet/pods/642d0ca0-2e0f-4b69-9484-a63d0a01f8a9/volumes" Jan 22 11:51:11 crc kubenswrapper[4874]: I0122 11:51:11.727636 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" event={"ID":"5937262c-c601-4bdc-9e86-06900a448320","Type":"ContainerStarted","Data":"af109f291683f6f23b12f6d5917d0dfbde6c90c5596bf42275ab78b1f661182c"} Jan 22 11:51:13 crc kubenswrapper[4874]: I0122 11:51:13.748496 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" event={"ID":"5937262c-c601-4bdc-9e86-06900a448320","Type":"ContainerStarted","Data":"23d543b70510b32616c6db1feceffde9b7debab80e4b5004506d534f966e35c1"} Jan 22 11:51:13 crc kubenswrapper[4874]: I0122 11:51:13.749127 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:13 crc kubenswrapper[4874]: I0122 11:51:13.790445 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:13 crc kubenswrapper[4874]: I0122 11:51:13.796485 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" podStartSLOduration=7.796461033 podStartE2EDuration="7.796461033s" podCreationTimestamp="2026-01-22 11:51:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:51:13.79350645 +0000 UTC m=+647.638577530" watchObservedRunningTime="2026-01-22 11:51:13.796461033 +0000 UTC m=+647.641532123" Jan 22 11:51:14 crc kubenswrapper[4874]: I0122 11:51:14.754374 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:14 crc kubenswrapper[4874]: I0122 11:51:14.754918 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:14 crc kubenswrapper[4874]: I0122 11:51:14.797065 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:51:19 crc kubenswrapper[4874]: I0122 11:51:19.716551 4874 scope.go:117] "RemoveContainer" containerID="55eeb9abd8c425711e374c107c22ec24d1741880327f226d7db5e06d67925630" Jan 22 11:51:19 crc kubenswrapper[4874]: E0122 11:51:19.716853 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-krrtc_openshift-multus(977746b5-ac1b-4b6e-bdbc-ddd90225e68c)\"" pod="openshift-multus/multus-krrtc" podUID="977746b5-ac1b-4b6e-bdbc-ddd90225e68c" Jan 22 11:51:27 crc kubenswrapper[4874]: I0122 11:51:27.140788 4874 scope.go:117] "RemoveContainer" containerID="cecfbaa0efaf8c435c3409ccad9deaa4cc25167f0b978622d1ab9c949c4024c8" Jan 22 11:51:27 crc kubenswrapper[4874]: I0122 11:51:27.835786 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/2.log" Jan 22 11:51:34 crc kubenswrapper[4874]: I0122 11:51:34.717072 4874 scope.go:117] "RemoveContainer" containerID="55eeb9abd8c425711e374c107c22ec24d1741880327f226d7db5e06d67925630" Jan 22 11:51:35 crc kubenswrapper[4874]: I0122 11:51:35.890021 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/2.log" Jan 22 11:51:35 crc kubenswrapper[4874]: I0122 11:51:35.892227 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-krrtc" event={"ID":"977746b5-ac1b-4b6e-bdbc-ddd90225e68c","Type":"ContainerStarted","Data":"c6cb636066ef9548b0f5770eab00a07ede7ca6a9963d9ed16cb71f112e67d961"} Jan 22 11:51:36 crc kubenswrapper[4874]: I0122 11:51:36.540102 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lbtk2" Jan 22 11:52:13 crc kubenswrapper[4874]: I0122 11:52:13.520765 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:52:13 crc kubenswrapper[4874]: I0122 11:52:13.521552 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.303384 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s7gc"] Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.304123 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5s7gc" podUID="37153c5d-6533-4973-99bd-f682b3c148d9" containerName="registry-server" containerID="cri-o://54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a" gracePeriod=30 Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.720877 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.881277 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-catalog-content\") pod \"37153c5d-6533-4973-99bd-f682b3c148d9\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.881481 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-utilities\") pod \"37153c5d-6533-4973-99bd-f682b3c148d9\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.881577 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slv5g\" (UniqueName: \"kubernetes.io/projected/37153c5d-6533-4973-99bd-f682b3c148d9-kube-api-access-slv5g\") pod \"37153c5d-6533-4973-99bd-f682b3c148d9\" (UID: \"37153c5d-6533-4973-99bd-f682b3c148d9\") " Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.883682 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-utilities" (OuterVolumeSpecName: "utilities") pod "37153c5d-6533-4973-99bd-f682b3c148d9" (UID: "37153c5d-6533-4973-99bd-f682b3c148d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.887294 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37153c5d-6533-4973-99bd-f682b3c148d9-kube-api-access-slv5g" (OuterVolumeSpecName: "kube-api-access-slv5g") pod "37153c5d-6533-4973-99bd-f682b3c148d9" (UID: "37153c5d-6533-4973-99bd-f682b3c148d9"). InnerVolumeSpecName "kube-api-access-slv5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.926580 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37153c5d-6533-4973-99bd-f682b3c148d9" (UID: "37153c5d-6533-4973-99bd-f682b3c148d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.983145 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slv5g\" (UniqueName: \"kubernetes.io/projected/37153c5d-6533-4973-99bd-f682b3c148d9-kube-api-access-slv5g\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.983186 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:17 crc kubenswrapper[4874]: I0122 11:52:17.983204 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37153c5d-6533-4973-99bd-f682b3c148d9-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.192531 4874 generic.go:334] "Generic (PLEG): container finished" podID="37153c5d-6533-4973-99bd-f682b3c148d9" containerID="54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a" exitCode=0 Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.192582 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s7gc" event={"ID":"37153c5d-6533-4973-99bd-f682b3c148d9","Type":"ContainerDied","Data":"54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a"} Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.192628 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s7gc" event={"ID":"37153c5d-6533-4973-99bd-f682b3c148d9","Type":"ContainerDied","Data":"3610f727eca8b59c0ef9ba7a4755e8243ae3a7e6694ae4e499109a32fd98ae07"} Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.192650 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s7gc" Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.192666 4874 scope.go:117] "RemoveContainer" containerID="54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a" Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.216690 4874 scope.go:117] "RemoveContainer" containerID="c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb" Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.246154 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s7gc"] Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.253999 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s7gc"] Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.254200 4874 scope.go:117] "RemoveContainer" containerID="684e002757e3dc3f4607d5fe73a314e71a31507700f72d618222be0e988a644f" Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.279004 4874 scope.go:117] "RemoveContainer" containerID="54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a" Jan 22 11:52:18 crc kubenswrapper[4874]: E0122 11:52:18.279655 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a\": container with ID starting with 54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a not found: ID does not exist" containerID="54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a" Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.279728 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a"} err="failed to get container status \"54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a\": rpc error: code = NotFound desc = could not find container \"54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a\": container with ID starting with 54db0b1ebec1433a5bd88c09690a6a8851c51a92a0a90d5576ee312fb1dc0d4a not found: ID does not exist" Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.279787 4874 scope.go:117] "RemoveContainer" containerID="c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb" Jan 22 11:52:18 crc kubenswrapper[4874]: E0122 11:52:18.280507 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb\": container with ID starting with c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb not found: ID does not exist" containerID="c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb" Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.280579 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb"} err="failed to get container status \"c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb\": rpc error: code = NotFound desc = could not find container \"c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb\": container with ID starting with c9c46af6c28993718e9a9e43b2e96b33eb0ad605dab2e2425b80f011941ca8fb not found: ID does not exist" Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.280617 4874 scope.go:117] "RemoveContainer" containerID="684e002757e3dc3f4607d5fe73a314e71a31507700f72d618222be0e988a644f" Jan 22 11:52:18 crc kubenswrapper[4874]: E0122 11:52:18.281171 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"684e002757e3dc3f4607d5fe73a314e71a31507700f72d618222be0e988a644f\": container with ID starting with 684e002757e3dc3f4607d5fe73a314e71a31507700f72d618222be0e988a644f not found: ID does not exist" containerID="684e002757e3dc3f4607d5fe73a314e71a31507700f72d618222be0e988a644f" Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.281222 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"684e002757e3dc3f4607d5fe73a314e71a31507700f72d618222be0e988a644f"} err="failed to get container status \"684e002757e3dc3f4607d5fe73a314e71a31507700f72d618222be0e988a644f\": rpc error: code = NotFound desc = could not find container \"684e002757e3dc3f4607d5fe73a314e71a31507700f72d618222be0e988a644f\": container with ID starting with 684e002757e3dc3f4607d5fe73a314e71a31507700f72d618222be0e988a644f not found: ID does not exist" Jan 22 11:52:18 crc kubenswrapper[4874]: I0122 11:52:18.727197 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37153c5d-6533-4973-99bd-f682b3c148d9" path="/var/lib/kubelet/pods/37153c5d-6533-4973-99bd-f682b3c148d9/volumes" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.531845 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk"] Jan 22 11:52:21 crc kubenswrapper[4874]: E0122 11:52:21.532377 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37153c5d-6533-4973-99bd-f682b3c148d9" containerName="registry-server" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.532427 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="37153c5d-6533-4973-99bd-f682b3c148d9" containerName="registry-server" Jan 22 11:52:21 crc kubenswrapper[4874]: E0122 11:52:21.532444 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37153c5d-6533-4973-99bd-f682b3c148d9" containerName="extract-utilities" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.532452 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="37153c5d-6533-4973-99bd-f682b3c148d9" containerName="extract-utilities" Jan 22 11:52:21 crc kubenswrapper[4874]: E0122 11:52:21.532470 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37153c5d-6533-4973-99bd-f682b3c148d9" containerName="extract-content" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.532478 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="37153c5d-6533-4973-99bd-f682b3c148d9" containerName="extract-content" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.532590 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="37153c5d-6533-4973-99bd-f682b3c148d9" containerName="registry-server" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.533430 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.538031 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.543446 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk"] Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.634573 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.634726 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.634800 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9v47\" (UniqueName: \"kubernetes.io/projected/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-kube-api-access-z9v47\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.736311 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9v47\" (UniqueName: \"kubernetes.io/projected/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-kube-api-access-z9v47\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.736631 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.736720 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.737125 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.737198 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.755458 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9v47\" (UniqueName: \"kubernetes.io/projected/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-kube-api-access-z9v47\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:21 crc kubenswrapper[4874]: I0122 11:52:21.847172 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:22 crc kubenswrapper[4874]: I0122 11:52:22.122625 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk"] Jan 22 11:52:22 crc kubenswrapper[4874]: W0122 11:52:22.141471 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9634e8c9_9571_45b6_ad7f_c7d68d40c75a.slice/crio-881c9c6a19afb1c0a047f5dade7d497915424b9f4b71441df5ffc1a8a17feeb6 WatchSource:0}: Error finding container 881c9c6a19afb1c0a047f5dade7d497915424b9f4b71441df5ffc1a8a17feeb6: Status 404 returned error can't find the container with id 881c9c6a19afb1c0a047f5dade7d497915424b9f4b71441df5ffc1a8a17feeb6 Jan 22 11:52:22 crc kubenswrapper[4874]: I0122 11:52:22.229243 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" event={"ID":"9634e8c9-9571-45b6-ad7f-c7d68d40c75a","Type":"ContainerStarted","Data":"881c9c6a19afb1c0a047f5dade7d497915424b9f4b71441df5ffc1a8a17feeb6"} Jan 22 11:52:23 crc kubenswrapper[4874]: I0122 11:52:23.238903 4874 generic.go:334] "Generic (PLEG): container finished" podID="9634e8c9-9571-45b6-ad7f-c7d68d40c75a" containerID="eae37d2395c60a97142ae4f46bc1135a93e0ddac46d0a371d1cc3b91beea07a3" exitCode=0 Jan 22 11:52:23 crc kubenswrapper[4874]: I0122 11:52:23.238987 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" event={"ID":"9634e8c9-9571-45b6-ad7f-c7d68d40c75a","Type":"ContainerDied","Data":"eae37d2395c60a97142ae4f46bc1135a93e0ddac46d0a371d1cc3b91beea07a3"} Jan 22 11:52:23 crc kubenswrapper[4874]: I0122 11:52:23.241946 4874 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 11:52:25 crc kubenswrapper[4874]: I0122 11:52:25.259317 4874 generic.go:334] "Generic (PLEG): container finished" podID="9634e8c9-9571-45b6-ad7f-c7d68d40c75a" containerID="14fc0f306c59c75bd148aa958b96e945154cf3e0099f2d9a38c28941880de481" exitCode=0 Jan 22 11:52:25 crc kubenswrapper[4874]: I0122 11:52:25.259415 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" event={"ID":"9634e8c9-9571-45b6-ad7f-c7d68d40c75a","Type":"ContainerDied","Data":"14fc0f306c59c75bd148aa958b96e945154cf3e0099f2d9a38c28941880de481"} Jan 22 11:52:26 crc kubenswrapper[4874]: I0122 11:52:26.272598 4874 generic.go:334] "Generic (PLEG): container finished" podID="9634e8c9-9571-45b6-ad7f-c7d68d40c75a" containerID="c1b49c9d548e2a106274d791e08e9815878bc897190996ea888b365bc8348c48" exitCode=0 Jan 22 11:52:26 crc kubenswrapper[4874]: I0122 11:52:26.272675 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" event={"ID":"9634e8c9-9571-45b6-ad7f-c7d68d40c75a","Type":"ContainerDied","Data":"c1b49c9d548e2a106274d791e08e9815878bc897190996ea888b365bc8348c48"} Jan 22 11:52:27 crc kubenswrapper[4874]: I0122 11:52:27.552522 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:27 crc kubenswrapper[4874]: I0122 11:52:27.724083 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-util\") pod \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " Jan 22 11:52:27 crc kubenswrapper[4874]: I0122 11:52:27.724275 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9v47\" (UniqueName: \"kubernetes.io/projected/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-kube-api-access-z9v47\") pod \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " Jan 22 11:52:27 crc kubenswrapper[4874]: I0122 11:52:27.724451 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-bundle\") pod \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\" (UID: \"9634e8c9-9571-45b6-ad7f-c7d68d40c75a\") " Jan 22 11:52:27 crc kubenswrapper[4874]: I0122 11:52:27.726192 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-bundle" (OuterVolumeSpecName: "bundle") pod "9634e8c9-9571-45b6-ad7f-c7d68d40c75a" (UID: "9634e8c9-9571-45b6-ad7f-c7d68d40c75a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:52:27 crc kubenswrapper[4874]: I0122 11:52:27.730092 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-kube-api-access-z9v47" (OuterVolumeSpecName: "kube-api-access-z9v47") pod "9634e8c9-9571-45b6-ad7f-c7d68d40c75a" (UID: "9634e8c9-9571-45b6-ad7f-c7d68d40c75a"). InnerVolumeSpecName "kube-api-access-z9v47". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:52:27 crc kubenswrapper[4874]: I0122 11:52:27.826811 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9v47\" (UniqueName: \"kubernetes.io/projected/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-kube-api-access-z9v47\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:27 crc kubenswrapper[4874]: I0122 11:52:27.826874 4874 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:27 crc kubenswrapper[4874]: I0122 11:52:27.980569 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-util" (OuterVolumeSpecName: "util") pod "9634e8c9-9571-45b6-ad7f-c7d68d40c75a" (UID: "9634e8c9-9571-45b6-ad7f-c7d68d40c75a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:52:28 crc kubenswrapper[4874]: I0122 11:52:28.030113 4874 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9634e8c9-9571-45b6-ad7f-c7d68d40c75a-util\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:28 crc kubenswrapper[4874]: I0122 11:52:28.289642 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" Jan 22 11:52:28 crc kubenswrapper[4874]: I0122 11:52:28.289638 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk" event={"ID":"9634e8c9-9571-45b6-ad7f-c7d68d40c75a","Type":"ContainerDied","Data":"881c9c6a19afb1c0a047f5dade7d497915424b9f4b71441df5ffc1a8a17feeb6"} Jan 22 11:52:28 crc kubenswrapper[4874]: I0122 11:52:28.289817 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="881c9c6a19afb1c0a047f5dade7d497915424b9f4b71441df5ffc1a8a17feeb6" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.524449 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7"] Jan 22 11:52:30 crc kubenswrapper[4874]: E0122 11:52:30.525288 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9634e8c9-9571-45b6-ad7f-c7d68d40c75a" containerName="extract" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.525355 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="9634e8c9-9571-45b6-ad7f-c7d68d40c75a" containerName="extract" Jan 22 11:52:30 crc kubenswrapper[4874]: E0122 11:52:30.525427 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9634e8c9-9571-45b6-ad7f-c7d68d40c75a" containerName="util" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.525484 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="9634e8c9-9571-45b6-ad7f-c7d68d40c75a" containerName="util" Jan 22 11:52:30 crc kubenswrapper[4874]: E0122 11:52:30.525540 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9634e8c9-9571-45b6-ad7f-c7d68d40c75a" containerName="pull" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.525590 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="9634e8c9-9571-45b6-ad7f-c7d68d40c75a" containerName="pull" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.525724 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="9634e8c9-9571-45b6-ad7f-c7d68d40c75a" containerName="extract" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.526415 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.529157 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.551698 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7"] Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.603081 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.603151 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.603330 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hrrh\" (UniqueName: \"kubernetes.io/projected/adad4abb-e416-4993-a166-0ac45ac75521-kube-api-access-8hrrh\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.704049 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.704100 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.704155 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hrrh\" (UniqueName: \"kubernetes.io/projected/adad4abb-e416-4993-a166-0ac45ac75521-kube-api-access-8hrrh\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.705058 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.705130 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.739068 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hrrh\" (UniqueName: \"kubernetes.io/projected/adad4abb-e416-4993-a166-0ac45ac75521-kube-api-access-8hrrh\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:30 crc kubenswrapper[4874]: I0122 11:52:30.843596 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.091659 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7"] Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.320864 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" event={"ID":"adad4abb-e416-4993-a166-0ac45ac75521","Type":"ContainerStarted","Data":"e2a672e474a2472d7f351a57371d8c07d2ea6ba5df72efe12375a43e4f3f5689"} Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.323365 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf"] Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.325304 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.331136 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf"] Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.413160 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4np4\" (UniqueName: \"kubernetes.io/projected/4db7503f-969b-4507-81bc-fb6ae0579495-kube-api-access-k4np4\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.413246 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.413335 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.514721 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4np4\" (UniqueName: \"kubernetes.io/projected/4db7503f-969b-4507-81bc-fb6ae0579495-kube-api-access-k4np4\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.514788 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.514832 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.515704 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.515714 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.547079 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4np4\" (UniqueName: \"kubernetes.io/projected/4db7503f-969b-4507-81bc-fb6ae0579495-kube-api-access-k4np4\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:31 crc kubenswrapper[4874]: I0122 11:52:31.640759 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:32 crc kubenswrapper[4874]: I0122 11:52:32.097268 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf"] Jan 22 11:52:32 crc kubenswrapper[4874]: I0122 11:52:32.332574 4874 generic.go:334] "Generic (PLEG): container finished" podID="adad4abb-e416-4993-a166-0ac45ac75521" containerID="80ace2274ec8740864fcc24a9c36c173227b492398f3b830bb5cf776977b640f" exitCode=0 Jan 22 11:52:32 crc kubenswrapper[4874]: I0122 11:52:32.333092 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" event={"ID":"adad4abb-e416-4993-a166-0ac45ac75521","Type":"ContainerDied","Data":"80ace2274ec8740864fcc24a9c36c173227b492398f3b830bb5cf776977b640f"} Jan 22 11:52:32 crc kubenswrapper[4874]: I0122 11:52:32.337690 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" event={"ID":"4db7503f-969b-4507-81bc-fb6ae0579495","Type":"ContainerStarted","Data":"9422670f2856c64b6f212b3001dace3d7fad6ff84246fe60c6d97aed2a878373"} Jan 22 11:52:33 crc kubenswrapper[4874]: I0122 11:52:33.349555 4874 generic.go:334] "Generic (PLEG): container finished" podID="4db7503f-969b-4507-81bc-fb6ae0579495" containerID="3e34f556369d896d056f66ff406929737cade6ea674c92a6bc257901a0bf77a7" exitCode=0 Jan 22 11:52:33 crc kubenswrapper[4874]: I0122 11:52:33.349597 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" event={"ID":"4db7503f-969b-4507-81bc-fb6ae0579495","Type":"ContainerDied","Data":"3e34f556369d896d056f66ff406929737cade6ea674c92a6bc257901a0bf77a7"} Jan 22 11:52:34 crc kubenswrapper[4874]: I0122 11:52:34.356980 4874 generic.go:334] "Generic (PLEG): container finished" podID="adad4abb-e416-4993-a166-0ac45ac75521" containerID="11fb58e2bc00eea5c08b47475d97aa9a4fd2f725b84421da88eb8c5594ad6ad3" exitCode=0 Jan 22 11:52:34 crc kubenswrapper[4874]: I0122 11:52:34.357100 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" event={"ID":"adad4abb-e416-4993-a166-0ac45ac75521","Type":"ContainerDied","Data":"11fb58e2bc00eea5c08b47475d97aa9a4fd2f725b84421da88eb8c5594ad6ad3"} Jan 22 11:52:35 crc kubenswrapper[4874]: I0122 11:52:35.369467 4874 generic.go:334] "Generic (PLEG): container finished" podID="adad4abb-e416-4993-a166-0ac45ac75521" containerID="ff4168d9b04656a0333f2d6281f8d25f2eccae29873f6c3a6e07cc5c3e2e8c25" exitCode=0 Jan 22 11:52:35 crc kubenswrapper[4874]: I0122 11:52:35.369592 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" event={"ID":"adad4abb-e416-4993-a166-0ac45ac75521","Type":"ContainerDied","Data":"ff4168d9b04656a0333f2d6281f8d25f2eccae29873f6c3a6e07cc5c3e2e8c25"} Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.376154 4874 generic.go:334] "Generic (PLEG): container finished" podID="4db7503f-969b-4507-81bc-fb6ae0579495" containerID="0134763dc613f6a61fcd3ffc52abae94a811a88ccffcae4fd175570f3b7d4852" exitCode=0 Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.377180 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" event={"ID":"4db7503f-969b-4507-81bc-fb6ae0579495","Type":"ContainerDied","Data":"0134763dc613f6a61fcd3ffc52abae94a811a88ccffcae4fd175570f3b7d4852"} Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.558671 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5"] Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.560260 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.575683 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5"] Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.604406 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.604445 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.604493 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6tsr\" (UniqueName: \"kubernetes.io/projected/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-kube-api-access-t6tsr\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.681772 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.705050 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6tsr\" (UniqueName: \"kubernetes.io/projected/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-kube-api-access-t6tsr\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.705124 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.705150 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.705590 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.706346 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.752372 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6tsr\" (UniqueName: \"kubernetes.io/projected/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-kube-api-access-t6tsr\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.805598 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-bundle\") pod \"adad4abb-e416-4993-a166-0ac45ac75521\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.805698 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hrrh\" (UniqueName: \"kubernetes.io/projected/adad4abb-e416-4993-a166-0ac45ac75521-kube-api-access-8hrrh\") pod \"adad4abb-e416-4993-a166-0ac45ac75521\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.805759 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-util\") pod \"adad4abb-e416-4993-a166-0ac45ac75521\" (UID: \"adad4abb-e416-4993-a166-0ac45ac75521\") " Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.806525 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-bundle" (OuterVolumeSpecName: "bundle") pod "adad4abb-e416-4993-a166-0ac45ac75521" (UID: "adad4abb-e416-4993-a166-0ac45ac75521"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.809019 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adad4abb-e416-4993-a166-0ac45ac75521-kube-api-access-8hrrh" (OuterVolumeSpecName: "kube-api-access-8hrrh") pod "adad4abb-e416-4993-a166-0ac45ac75521" (UID: "adad4abb-e416-4993-a166-0ac45ac75521"). InnerVolumeSpecName "kube-api-access-8hrrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.823456 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-util" (OuterVolumeSpecName: "util") pod "adad4abb-e416-4993-a166-0ac45ac75521" (UID: "adad4abb-e416-4993-a166-0ac45ac75521"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.890453 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.907206 4874 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-util\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.907235 4874 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/adad4abb-e416-4993-a166-0ac45ac75521-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:36 crc kubenswrapper[4874]: I0122 11:52:36.907244 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hrrh\" (UniqueName: \"kubernetes.io/projected/adad4abb-e416-4993-a166-0ac45ac75521-kube-api-access-8hrrh\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:37 crc kubenswrapper[4874]: I0122 11:52:37.158371 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5"] Jan 22 11:52:37 crc kubenswrapper[4874]: W0122 11:52:37.168634 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a48ea4e_8d83_4484_8cab_e9a38e86a2e1.slice/crio-3052c644abdb9c0b9e56a9a8e5229cfd6a7be0945646a4a78aa60b838b8c13d8 WatchSource:0}: Error finding container 3052c644abdb9c0b9e56a9a8e5229cfd6a7be0945646a4a78aa60b838b8c13d8: Status 404 returned error can't find the container with id 3052c644abdb9c0b9e56a9a8e5229cfd6a7be0945646a4a78aa60b838b8c13d8 Jan 22 11:52:37 crc kubenswrapper[4874]: I0122 11:52:37.382824 4874 generic.go:334] "Generic (PLEG): container finished" podID="4db7503f-969b-4507-81bc-fb6ae0579495" containerID="0d803dc33d5b1561687c54af4a3c7e656bf69906e6220e94080cd3625f5561bb" exitCode=0 Jan 22 11:52:37 crc kubenswrapper[4874]: I0122 11:52:37.382876 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" event={"ID":"4db7503f-969b-4507-81bc-fb6ae0579495","Type":"ContainerDied","Data":"0d803dc33d5b1561687c54af4a3c7e656bf69906e6220e94080cd3625f5561bb"} Jan 22 11:52:37 crc kubenswrapper[4874]: I0122 11:52:37.384755 4874 generic.go:334] "Generic (PLEG): container finished" podID="9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" containerID="26ee57824a8248b7caf445e3d93c945e9156e940ffa5f1516e4ab1cefcd20ae4" exitCode=0 Jan 22 11:52:37 crc kubenswrapper[4874]: I0122 11:52:37.384842 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" event={"ID":"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1","Type":"ContainerDied","Data":"26ee57824a8248b7caf445e3d93c945e9156e940ffa5f1516e4ab1cefcd20ae4"} Jan 22 11:52:37 crc kubenswrapper[4874]: I0122 11:52:37.384897 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" event={"ID":"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1","Type":"ContainerStarted","Data":"3052c644abdb9c0b9e56a9a8e5229cfd6a7be0945646a4a78aa60b838b8c13d8"} Jan 22 11:52:37 crc kubenswrapper[4874]: I0122 11:52:37.388810 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" event={"ID":"adad4abb-e416-4993-a166-0ac45ac75521","Type":"ContainerDied","Data":"e2a672e474a2472d7f351a57371d8c07d2ea6ba5df72efe12375a43e4f3f5689"} Jan 22 11:52:37 crc kubenswrapper[4874]: I0122 11:52:37.388843 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2a672e474a2472d7f351a57371d8c07d2ea6ba5df72efe12375a43e4f3f5689" Jan 22 11:52:37 crc kubenswrapper[4874]: I0122 11:52:37.388897 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7" Jan 22 11:52:38 crc kubenswrapper[4874]: I0122 11:52:38.602818 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:38 crc kubenswrapper[4874]: I0122 11:52:38.728429 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-bundle\") pod \"4db7503f-969b-4507-81bc-fb6ae0579495\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " Jan 22 11:52:38 crc kubenswrapper[4874]: I0122 11:52:38.730020 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-util\") pod \"4db7503f-969b-4507-81bc-fb6ae0579495\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " Jan 22 11:52:38 crc kubenswrapper[4874]: I0122 11:52:38.730170 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4np4\" (UniqueName: \"kubernetes.io/projected/4db7503f-969b-4507-81bc-fb6ae0579495-kube-api-access-k4np4\") pod \"4db7503f-969b-4507-81bc-fb6ae0579495\" (UID: \"4db7503f-969b-4507-81bc-fb6ae0579495\") " Jan 22 11:52:38 crc kubenswrapper[4874]: I0122 11:52:38.730984 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-bundle" (OuterVolumeSpecName: "bundle") pod "4db7503f-969b-4507-81bc-fb6ae0579495" (UID: "4db7503f-969b-4507-81bc-fb6ae0579495"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:52:38 crc kubenswrapper[4874]: I0122 11:52:38.731469 4874 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:38 crc kubenswrapper[4874]: I0122 11:52:38.748636 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4db7503f-969b-4507-81bc-fb6ae0579495-kube-api-access-k4np4" (OuterVolumeSpecName: "kube-api-access-k4np4") pod "4db7503f-969b-4507-81bc-fb6ae0579495" (UID: "4db7503f-969b-4507-81bc-fb6ae0579495"). InnerVolumeSpecName "kube-api-access-k4np4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:52:38 crc kubenswrapper[4874]: I0122 11:52:38.752665 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-util" (OuterVolumeSpecName: "util") pod "4db7503f-969b-4507-81bc-fb6ae0579495" (UID: "4db7503f-969b-4507-81bc-fb6ae0579495"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:52:38 crc kubenswrapper[4874]: I0122 11:52:38.832346 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4np4\" (UniqueName: \"kubernetes.io/projected/4db7503f-969b-4507-81bc-fb6ae0579495-kube-api-access-k4np4\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:38 crc kubenswrapper[4874]: I0122 11:52:38.832407 4874 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4db7503f-969b-4507-81bc-fb6ae0579495-util\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:39 crc kubenswrapper[4874]: I0122 11:52:39.406468 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" event={"ID":"4db7503f-969b-4507-81bc-fb6ae0579495","Type":"ContainerDied","Data":"9422670f2856c64b6f212b3001dace3d7fad6ff84246fe60c6d97aed2a878373"} Jan 22 11:52:39 crc kubenswrapper[4874]: I0122 11:52:39.406508 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9422670f2856c64b6f212b3001dace3d7fad6ff84246fe60c6d97aed2a878373" Jan 22 11:52:39 crc kubenswrapper[4874]: I0122 11:52:39.406529 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.221385 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-w7v89"] Jan 22 11:52:40 crc kubenswrapper[4874]: E0122 11:52:40.221651 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adad4abb-e416-4993-a166-0ac45ac75521" containerName="pull" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.221668 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="adad4abb-e416-4993-a166-0ac45ac75521" containerName="pull" Jan 22 11:52:40 crc kubenswrapper[4874]: E0122 11:52:40.221685 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adad4abb-e416-4993-a166-0ac45ac75521" containerName="extract" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.221692 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="adad4abb-e416-4993-a166-0ac45ac75521" containerName="extract" Jan 22 11:52:40 crc kubenswrapper[4874]: E0122 11:52:40.221702 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adad4abb-e416-4993-a166-0ac45ac75521" containerName="util" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.221710 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="adad4abb-e416-4993-a166-0ac45ac75521" containerName="util" Jan 22 11:52:40 crc kubenswrapper[4874]: E0122 11:52:40.221720 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db7503f-969b-4507-81bc-fb6ae0579495" containerName="pull" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.221727 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db7503f-969b-4507-81bc-fb6ae0579495" containerName="pull" Jan 22 11:52:40 crc kubenswrapper[4874]: E0122 11:52:40.221738 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db7503f-969b-4507-81bc-fb6ae0579495" containerName="util" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.221744 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db7503f-969b-4507-81bc-fb6ae0579495" containerName="util" Jan 22 11:52:40 crc kubenswrapper[4874]: E0122 11:52:40.221760 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db7503f-969b-4507-81bc-fb6ae0579495" containerName="extract" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.221767 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db7503f-969b-4507-81bc-fb6ae0579495" containerName="extract" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.221898 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="4db7503f-969b-4507-81bc-fb6ae0579495" containerName="extract" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.221907 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="adad4abb-e416-4993-a166-0ac45ac75521" containerName="extract" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.222307 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w7v89" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.224387 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.225017 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-7hlw4" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.225207 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.233378 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-w7v89"] Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.355048 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzc57\" (UniqueName: \"kubernetes.io/projected/71a74c6d-7ff3-4a3a-8c9d-bfa212b381b2-kube-api-access-lzc57\") pod \"obo-prometheus-operator-68bc856cb9-w7v89\" (UID: \"71a74c6d-7ff3-4a3a-8c9d-bfa212b381b2\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w7v89" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.393444 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8"] Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.394184 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.396384 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.396790 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-blzqr" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.398494 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr"] Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.399237 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.407902 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr"] Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.411306 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8"] Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.456797 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzc57\" (UniqueName: \"kubernetes.io/projected/71a74c6d-7ff3-4a3a-8c9d-bfa212b381b2-kube-api-access-lzc57\") pod \"obo-prometheus-operator-68bc856cb9-w7v89\" (UID: \"71a74c6d-7ff3-4a3a-8c9d-bfa212b381b2\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w7v89" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.474832 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzc57\" (UniqueName: \"kubernetes.io/projected/71a74c6d-7ff3-4a3a-8c9d-bfa212b381b2-kube-api-access-lzc57\") pod \"obo-prometheus-operator-68bc856cb9-w7v89\" (UID: \"71a74c6d-7ff3-4a3a-8c9d-bfa212b381b2\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w7v89" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.547469 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w7v89" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.558124 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7a18bf90-2af7-4296-9416-1368e89a8a03-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr\" (UID: \"7a18bf90-2af7-4296-9416-1368e89a8a03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.558186 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7a18bf90-2af7-4296-9416-1368e89a8a03-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr\" (UID: \"7a18bf90-2af7-4296-9416-1368e89a8a03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.558293 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60c00b52-d6b8-440e-9e60-966b44a87577-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8\" (UID: \"60c00b52-d6b8-440e-9e60-966b44a87577\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.558342 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60c00b52-d6b8-440e-9e60-966b44a87577-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8\" (UID: \"60c00b52-d6b8-440e-9e60-966b44a87577\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.573329 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-jblsg"] Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.574067 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-jblsg" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.575867 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-9msn9" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.576934 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.585314 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-jblsg"] Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.660066 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7a18bf90-2af7-4296-9416-1368e89a8a03-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr\" (UID: \"7a18bf90-2af7-4296-9416-1368e89a8a03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.660157 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60c00b52-d6b8-440e-9e60-966b44a87577-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8\" (UID: \"60c00b52-d6b8-440e-9e60-966b44a87577\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.660208 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60c00b52-d6b8-440e-9e60-966b44a87577-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8\" (UID: \"60c00b52-d6b8-440e-9e60-966b44a87577\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.660255 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7a18bf90-2af7-4296-9416-1368e89a8a03-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr\" (UID: \"7a18bf90-2af7-4296-9416-1368e89a8a03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.664718 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7a18bf90-2af7-4296-9416-1368e89a8a03-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr\" (UID: \"7a18bf90-2af7-4296-9416-1368e89a8a03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.665796 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7a18bf90-2af7-4296-9416-1368e89a8a03-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr\" (UID: \"7a18bf90-2af7-4296-9416-1368e89a8a03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.666714 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-kqrkd"] Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.667518 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.669147 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60c00b52-d6b8-440e-9e60-966b44a87577-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8\" (UID: \"60c00b52-d6b8-440e-9e60-966b44a87577\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.669159 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60c00b52-d6b8-440e-9e60-966b44a87577-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8\" (UID: \"60c00b52-d6b8-440e-9e60-966b44a87577\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.669448 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-jpmjt" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.684305 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-kqrkd"] Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.708996 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.728030 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.761671 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/edd7744e-1336-486a-90de-6568bae7f788-observability-operator-tls\") pod \"observability-operator-59bdc8b94-jblsg\" (UID: \"edd7744e-1336-486a-90de-6568bae7f788\") " pod="openshift-operators/observability-operator-59bdc8b94-jblsg" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.761735 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgkks\" (UniqueName: \"kubernetes.io/projected/edd7744e-1336-486a-90de-6568bae7f788-kube-api-access-mgkks\") pod \"observability-operator-59bdc8b94-jblsg\" (UID: \"edd7744e-1336-486a-90de-6568bae7f788\") " pod="openshift-operators/observability-operator-59bdc8b94-jblsg" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.862549 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/97207973-f6a4-4068-b777-c964c12092fd-openshift-service-ca\") pod \"perses-operator-5bf474d74f-kqrkd\" (UID: \"97207973-f6a4-4068-b777-c964c12092fd\") " pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.862884 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgkks\" (UniqueName: \"kubernetes.io/projected/edd7744e-1336-486a-90de-6568bae7f788-kube-api-access-mgkks\") pod \"observability-operator-59bdc8b94-jblsg\" (UID: \"edd7744e-1336-486a-90de-6568bae7f788\") " pod="openshift-operators/observability-operator-59bdc8b94-jblsg" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.862923 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n25b\" (UniqueName: \"kubernetes.io/projected/97207973-f6a4-4068-b777-c964c12092fd-kube-api-access-4n25b\") pod \"perses-operator-5bf474d74f-kqrkd\" (UID: \"97207973-f6a4-4068-b777-c964c12092fd\") " pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.862979 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/edd7744e-1336-486a-90de-6568bae7f788-observability-operator-tls\") pod \"observability-operator-59bdc8b94-jblsg\" (UID: \"edd7744e-1336-486a-90de-6568bae7f788\") " pod="openshift-operators/observability-operator-59bdc8b94-jblsg" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.874278 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/edd7744e-1336-486a-90de-6568bae7f788-observability-operator-tls\") pod \"observability-operator-59bdc8b94-jblsg\" (UID: \"edd7744e-1336-486a-90de-6568bae7f788\") " pod="openshift-operators/observability-operator-59bdc8b94-jblsg" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.878251 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgkks\" (UniqueName: \"kubernetes.io/projected/edd7744e-1336-486a-90de-6568bae7f788-kube-api-access-mgkks\") pod \"observability-operator-59bdc8b94-jblsg\" (UID: \"edd7744e-1336-486a-90de-6568bae7f788\") " pod="openshift-operators/observability-operator-59bdc8b94-jblsg" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.897762 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-jblsg" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.964748 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/97207973-f6a4-4068-b777-c964c12092fd-openshift-service-ca\") pod \"perses-operator-5bf474d74f-kqrkd\" (UID: \"97207973-f6a4-4068-b777-c964c12092fd\") " pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.964829 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n25b\" (UniqueName: \"kubernetes.io/projected/97207973-f6a4-4068-b777-c964c12092fd-kube-api-access-4n25b\") pod \"perses-operator-5bf474d74f-kqrkd\" (UID: \"97207973-f6a4-4068-b777-c964c12092fd\") " pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.966147 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/97207973-f6a4-4068-b777-c964c12092fd-openshift-service-ca\") pod \"perses-operator-5bf474d74f-kqrkd\" (UID: \"97207973-f6a4-4068-b777-c964c12092fd\") " pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" Jan 22 11:52:40 crc kubenswrapper[4874]: I0122 11:52:40.980845 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n25b\" (UniqueName: \"kubernetes.io/projected/97207973-f6a4-4068-b777-c964c12092fd-kube-api-access-4n25b\") pod \"perses-operator-5bf474d74f-kqrkd\" (UID: \"97207973-f6a4-4068-b777-c964c12092fd\") " pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" Jan 22 11:52:41 crc kubenswrapper[4874]: I0122 11:52:41.034494 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" Jan 22 11:52:42 crc kubenswrapper[4874]: I0122 11:52:42.775912 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr"] Jan 22 11:52:42 crc kubenswrapper[4874]: I0122 11:52:42.907047 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-w7v89"] Jan 22 11:52:42 crc kubenswrapper[4874]: I0122 11:52:42.963808 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-kqrkd"] Jan 22 11:52:43 crc kubenswrapper[4874]: I0122 11:52:43.009188 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8"] Jan 22 11:52:43 crc kubenswrapper[4874]: W0122 11:52:43.020697 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60c00b52_d6b8_440e_9e60_966b44a87577.slice/crio-071374853f0b22bfa420a0207169cb3856cdc1d32c1f0b80507f25cd82952c66 WatchSource:0}: Error finding container 071374853f0b22bfa420a0207169cb3856cdc1d32c1f0b80507f25cd82952c66: Status 404 returned error can't find the container with id 071374853f0b22bfa420a0207169cb3856cdc1d32c1f0b80507f25cd82952c66 Jan 22 11:52:43 crc kubenswrapper[4874]: I0122 11:52:43.053008 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-jblsg"] Jan 22 11:52:43 crc kubenswrapper[4874]: W0122 11:52:43.059814 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedd7744e_1336_486a_90de_6568bae7f788.slice/crio-84faddd9561c05f9272a73c715714a774feee2a07caebf106d514f9188949bec WatchSource:0}: Error finding container 84faddd9561c05f9272a73c715714a774feee2a07caebf106d514f9188949bec: Status 404 returned error can't find the container with id 84faddd9561c05f9272a73c715714a774feee2a07caebf106d514f9188949bec Jan 22 11:52:43 crc kubenswrapper[4874]: I0122 11:52:43.425999 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" event={"ID":"97207973-f6a4-4068-b777-c964c12092fd","Type":"ContainerStarted","Data":"af7d56521f22776d7a26cbea5573c03beeed3c46a601ea393e183b076b52007a"} Jan 22 11:52:43 crc kubenswrapper[4874]: I0122 11:52:43.427116 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w7v89" event={"ID":"71a74c6d-7ff3-4a3a-8c9d-bfa212b381b2","Type":"ContainerStarted","Data":"2939ac2a943208155d80f64dba9c1fe0e1633524cc7581fc1b9217a8fb92bff3"} Jan 22 11:52:43 crc kubenswrapper[4874]: I0122 11:52:43.428177 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8" event={"ID":"60c00b52-d6b8-440e-9e60-966b44a87577","Type":"ContainerStarted","Data":"071374853f0b22bfa420a0207169cb3856cdc1d32c1f0b80507f25cd82952c66"} Jan 22 11:52:43 crc kubenswrapper[4874]: I0122 11:52:43.429656 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-jblsg" event={"ID":"edd7744e-1336-486a-90de-6568bae7f788","Type":"ContainerStarted","Data":"84faddd9561c05f9272a73c715714a774feee2a07caebf106d514f9188949bec"} Jan 22 11:52:43 crc kubenswrapper[4874]: I0122 11:52:43.431093 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr" event={"ID":"7a18bf90-2af7-4296-9416-1368e89a8a03","Type":"ContainerStarted","Data":"ee76d877051ccfd2790773e2f151b9634ad8539a8628521bb17f302f5b024389"} Jan 22 11:52:43 crc kubenswrapper[4874]: I0122 11:52:43.520438 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:52:43 crc kubenswrapper[4874]: I0122 11:52:43.520815 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:52:45 crc kubenswrapper[4874]: I0122 11:52:45.446645 4874 generic.go:334] "Generic (PLEG): container finished" podID="9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" containerID="8f4fba73edd5997fdd330668ad51504fc6f4795841dfda0808d04ee3f53a0555" exitCode=0 Jan 22 11:52:45 crc kubenswrapper[4874]: I0122 11:52:45.446830 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" event={"ID":"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1","Type":"ContainerDied","Data":"8f4fba73edd5997fdd330668ad51504fc6f4795841dfda0808d04ee3f53a0555"} Jan 22 11:52:46 crc kubenswrapper[4874]: I0122 11:52:46.457128 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" event={"ID":"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1","Type":"ContainerStarted","Data":"fc638188b36fb44914ecdf386b1c4e8e730c00e385e24ba36def328ef5b5bbcb"} Jan 22 11:52:46 crc kubenswrapper[4874]: I0122 11:52:46.481119 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" podStartSLOduration=3.077951051 podStartE2EDuration="10.481091941s" podCreationTimestamp="2026-01-22 11:52:36 +0000 UTC" firstStartedPulling="2026-01-22 11:52:37.387343059 +0000 UTC m=+731.232414129" lastFinishedPulling="2026-01-22 11:52:44.790483949 +0000 UTC m=+738.635555019" observedRunningTime="2026-01-22 11:52:46.477103278 +0000 UTC m=+740.322174368" watchObservedRunningTime="2026-01-22 11:52:46.481091941 +0000 UTC m=+740.326163031" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.291690 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-864f7dd768-szq44"] Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.292562 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-864f7dd768-szq44" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.296831 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.297114 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.297281 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-gmlh6" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.297503 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.322258 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-864f7dd768-szq44"] Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.458575 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b54bc0d4-cf91-46af-bc4b-ea963cbd59bf-webhook-cert\") pod \"elastic-operator-864f7dd768-szq44\" (UID: \"b54bc0d4-cf91-46af-bc4b-ea963cbd59bf\") " pod="service-telemetry/elastic-operator-864f7dd768-szq44" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.458673 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b54bc0d4-cf91-46af-bc4b-ea963cbd59bf-apiservice-cert\") pod \"elastic-operator-864f7dd768-szq44\" (UID: \"b54bc0d4-cf91-46af-bc4b-ea963cbd59bf\") " pod="service-telemetry/elastic-operator-864f7dd768-szq44" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.458768 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db6l9\" (UniqueName: \"kubernetes.io/projected/b54bc0d4-cf91-46af-bc4b-ea963cbd59bf-kube-api-access-db6l9\") pod \"elastic-operator-864f7dd768-szq44\" (UID: \"b54bc0d4-cf91-46af-bc4b-ea963cbd59bf\") " pod="service-telemetry/elastic-operator-864f7dd768-szq44" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.472435 4874 generic.go:334] "Generic (PLEG): container finished" podID="9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" containerID="fc638188b36fb44914ecdf386b1c4e8e730c00e385e24ba36def328ef5b5bbcb" exitCode=0 Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.472485 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" event={"ID":"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1","Type":"ContainerDied","Data":"fc638188b36fb44914ecdf386b1c4e8e730c00e385e24ba36def328ef5b5bbcb"} Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.560011 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db6l9\" (UniqueName: \"kubernetes.io/projected/b54bc0d4-cf91-46af-bc4b-ea963cbd59bf-kube-api-access-db6l9\") pod \"elastic-operator-864f7dd768-szq44\" (UID: \"b54bc0d4-cf91-46af-bc4b-ea963cbd59bf\") " pod="service-telemetry/elastic-operator-864f7dd768-szq44" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.560905 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b54bc0d4-cf91-46af-bc4b-ea963cbd59bf-webhook-cert\") pod \"elastic-operator-864f7dd768-szq44\" (UID: \"b54bc0d4-cf91-46af-bc4b-ea963cbd59bf\") " pod="service-telemetry/elastic-operator-864f7dd768-szq44" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.560983 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b54bc0d4-cf91-46af-bc4b-ea963cbd59bf-apiservice-cert\") pod \"elastic-operator-864f7dd768-szq44\" (UID: \"b54bc0d4-cf91-46af-bc4b-ea963cbd59bf\") " pod="service-telemetry/elastic-operator-864f7dd768-szq44" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.573492 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b54bc0d4-cf91-46af-bc4b-ea963cbd59bf-webhook-cert\") pod \"elastic-operator-864f7dd768-szq44\" (UID: \"b54bc0d4-cf91-46af-bc4b-ea963cbd59bf\") " pod="service-telemetry/elastic-operator-864f7dd768-szq44" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.581260 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db6l9\" (UniqueName: \"kubernetes.io/projected/b54bc0d4-cf91-46af-bc4b-ea963cbd59bf-kube-api-access-db6l9\") pod \"elastic-operator-864f7dd768-szq44\" (UID: \"b54bc0d4-cf91-46af-bc4b-ea963cbd59bf\") " pod="service-telemetry/elastic-operator-864f7dd768-szq44" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.582255 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b54bc0d4-cf91-46af-bc4b-ea963cbd59bf-apiservice-cert\") pod \"elastic-operator-864f7dd768-szq44\" (UID: \"b54bc0d4-cf91-46af-bc4b-ea963cbd59bf\") " pod="service-telemetry/elastic-operator-864f7dd768-szq44" Jan 22 11:52:47 crc kubenswrapper[4874]: I0122 11:52:47.628829 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-864f7dd768-szq44" Jan 22 11:52:48 crc kubenswrapper[4874]: I0122 11:52:48.152229 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-864f7dd768-szq44"] Jan 22 11:52:48 crc kubenswrapper[4874]: I0122 11:52:48.484124 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-864f7dd768-szq44" event={"ID":"b54bc0d4-cf91-46af-bc4b-ea963cbd59bf","Type":"ContainerStarted","Data":"2a3fc82ab4d3e0219b64906ab76f1bf5d73445b7dc86b19fab025743848cb272"} Jan 22 11:52:48 crc kubenswrapper[4874]: I0122 11:52:48.986180 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.100789 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6tsr\" (UniqueName: \"kubernetes.io/projected/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-kube-api-access-t6tsr\") pod \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.101114 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-util\") pod \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.101145 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-bundle\") pod \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\" (UID: \"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1\") " Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.102289 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-bundle" (OuterVolumeSpecName: "bundle") pod "9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" (UID: "9a48ea4e-8d83-4484-8cab-e9a38e86a2e1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.106216 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-kube-api-access-t6tsr" (OuterVolumeSpecName: "kube-api-access-t6tsr") pod "9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" (UID: "9a48ea4e-8d83-4484-8cab-e9a38e86a2e1"). InnerVolumeSpecName "kube-api-access-t6tsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.112202 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-util" (OuterVolumeSpecName: "util") pod "9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" (UID: "9a48ea4e-8d83-4484-8cab-e9a38e86a2e1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.202588 4874 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-util\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.202620 4874 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.202629 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6tsr\" (UniqueName: \"kubernetes.io/projected/9a48ea4e-8d83-4484-8cab-e9a38e86a2e1-kube-api-access-t6tsr\") on node \"crc\" DevicePath \"\"" Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.492245 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" event={"ID":"9a48ea4e-8d83-4484-8cab-e9a38e86a2e1","Type":"ContainerDied","Data":"3052c644abdb9c0b9e56a9a8e5229cfd6a7be0945646a4a78aa60b838b8c13d8"} Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.492278 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3052c644abdb9c0b9e56a9a8e5229cfd6a7be0945646a4a78aa60b838b8c13d8" Jan 22 11:52:49 crc kubenswrapper[4874]: I0122 11:52:49.492331 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5" Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.062673 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-p9brk"] Jan 22 11:52:50 crc kubenswrapper[4874]: E0122 11:52:50.062878 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" containerName="extract" Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.062890 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" containerName="extract" Jan 22 11:52:50 crc kubenswrapper[4874]: E0122 11:52:50.062907 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" containerName="util" Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.062913 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" containerName="util" Jan 22 11:52:50 crc kubenswrapper[4874]: E0122 11:52:50.062921 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" containerName="pull" Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.062928 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" containerName="pull" Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.063018 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a48ea4e-8d83-4484-8cab-e9a38e86a2e1" containerName="extract" Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.063363 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-p9brk" Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.067536 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-kgzql" Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.083431 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-p9brk"] Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.129103 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhpmr\" (UniqueName: \"kubernetes.io/projected/cd733750-1915-4813-9e44-ec3777ce9c53-kube-api-access-vhpmr\") pod \"interconnect-operator-5bb49f789d-p9brk\" (UID: \"cd733750-1915-4813-9e44-ec3777ce9c53\") " pod="service-telemetry/interconnect-operator-5bb49f789d-p9brk" Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.230222 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhpmr\" (UniqueName: \"kubernetes.io/projected/cd733750-1915-4813-9e44-ec3777ce9c53-kube-api-access-vhpmr\") pod \"interconnect-operator-5bb49f789d-p9brk\" (UID: \"cd733750-1915-4813-9e44-ec3777ce9c53\") " pod="service-telemetry/interconnect-operator-5bb49f789d-p9brk" Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.265606 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhpmr\" (UniqueName: \"kubernetes.io/projected/cd733750-1915-4813-9e44-ec3777ce9c53-kube-api-access-vhpmr\") pod \"interconnect-operator-5bb49f789d-p9brk\" (UID: \"cd733750-1915-4813-9e44-ec3777ce9c53\") " pod="service-telemetry/interconnect-operator-5bb49f789d-p9brk" Jan 22 11:52:50 crc kubenswrapper[4874]: I0122 11:52:50.374364 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-p9brk" Jan 22 11:52:55 crc kubenswrapper[4874]: I0122 11:52:55.502983 4874 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 22 11:52:59 crc kubenswrapper[4874]: W0122 11:52:59.095784 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd733750_1915_4813_9e44_ec3777ce9c53.slice/crio-1d443ee2c729b02d94ef1b22ae25a71170a0d4e111425d5991a0bbde175c0210 WatchSource:0}: Error finding container 1d443ee2c729b02d94ef1b22ae25a71170a0d4e111425d5991a0bbde175c0210: Status 404 returned error can't find the container with id 1d443ee2c729b02d94ef1b22ae25a71170a0d4e111425d5991a0bbde175c0210 Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.095826 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-p9brk"] Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.574313 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" event={"ID":"97207973-f6a4-4068-b777-c964c12092fd","Type":"ContainerStarted","Data":"2d721160c38f80fe5e62eb37a3c92e1d8f8a21b4cd2a62f552f221e4ea494537"} Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.574588 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.576272 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w7v89" event={"ID":"71a74c6d-7ff3-4a3a-8c9d-bfa212b381b2","Type":"ContainerStarted","Data":"97749ab37a11c6d60882b8081be341755317bf7f62aeec71115ed701a862b8d5"} Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.577752 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8" event={"ID":"60c00b52-d6b8-440e-9e60-966b44a87577","Type":"ContainerStarted","Data":"ace16ec90ba20ee4aa51e800b37f8468cb0eda193e34cfc51a6f3ab39f0f8ea3"} Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.579233 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-p9brk" event={"ID":"cd733750-1915-4813-9e44-ec3777ce9c53","Type":"ContainerStarted","Data":"1d443ee2c729b02d94ef1b22ae25a71170a0d4e111425d5991a0bbde175c0210"} Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.580801 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-864f7dd768-szq44" event={"ID":"b54bc0d4-cf91-46af-bc4b-ea963cbd59bf","Type":"ContainerStarted","Data":"8b257cf2404e4806872fd315593bafd9a67855ffb0ab16f030c72d3a7298a7cb"} Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.582855 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-jblsg" event={"ID":"edd7744e-1336-486a-90de-6568bae7f788","Type":"ContainerStarted","Data":"82c9ab721e7f28c9a44ad0007f0e88d66266eda8f1cc1ee9d707685d33737fbd"} Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.583048 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-jblsg" Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.584469 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr" event={"ID":"7a18bf90-2af7-4296-9416-1368e89a8a03","Type":"ContainerStarted","Data":"3081848caeb9b63e9b89af3d3c8d8dfb432a0f37c0ba57d07f79e3b89fd37bd6"} Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.586153 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-jblsg" Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.599447 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" podStartSLOduration=4.000512991 podStartE2EDuration="19.59943077s" podCreationTimestamp="2026-01-22 11:52:40 +0000 UTC" firstStartedPulling="2026-01-22 11:52:42.979897693 +0000 UTC m=+736.824968763" lastFinishedPulling="2026-01-22 11:52:58.578815472 +0000 UTC m=+752.423886542" observedRunningTime="2026-01-22 11:52:59.59652729 +0000 UTC m=+753.441598370" watchObservedRunningTime="2026-01-22 11:52:59.59943077 +0000 UTC m=+753.444501850" Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.623555 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr" podStartSLOduration=3.91816611 podStartE2EDuration="19.623533495s" podCreationTimestamp="2026-01-22 11:52:40 +0000 UTC" firstStartedPulling="2026-01-22 11:52:42.805558241 +0000 UTC m=+736.650629321" lastFinishedPulling="2026-01-22 11:52:58.510925636 +0000 UTC m=+752.355996706" observedRunningTime="2026-01-22 11:52:59.620714418 +0000 UTC m=+753.465785498" watchObservedRunningTime="2026-01-22 11:52:59.623533495 +0000 UTC m=+753.468604575" Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.671175 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-jblsg" podStartSLOduration=4.060187264 podStartE2EDuration="19.671156495s" podCreationTimestamp="2026-01-22 11:52:40 +0000 UTC" firstStartedPulling="2026-01-22 11:52:43.067570859 +0000 UTC m=+736.912641929" lastFinishedPulling="2026-01-22 11:52:58.67854009 +0000 UTC m=+752.523611160" observedRunningTime="2026-01-22 11:52:59.669894975 +0000 UTC m=+753.514966045" watchObservedRunningTime="2026-01-22 11:52:59.671156495 +0000 UTC m=+753.516227565" Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.676317 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w7v89" podStartSLOduration=4.021122368 podStartE2EDuration="19.676302764s" podCreationTimestamp="2026-01-22 11:52:40 +0000 UTC" firstStartedPulling="2026-01-22 11:52:42.922614854 +0000 UTC m=+736.767685924" lastFinishedPulling="2026-01-22 11:52:58.57779526 +0000 UTC m=+752.422866320" observedRunningTime="2026-01-22 11:52:59.644593694 +0000 UTC m=+753.489664784" watchObservedRunningTime="2026-01-22 11:52:59.676302764 +0000 UTC m=+753.521373834" Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.705046 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-864f7dd768-szq44" podStartSLOduration=2.283377763 podStartE2EDuration="12.70502919s" podCreationTimestamp="2026-01-22 11:52:47 +0000 UTC" firstStartedPulling="2026-01-22 11:52:48.188360658 +0000 UTC m=+742.033431728" lastFinishedPulling="2026-01-22 11:52:58.610012085 +0000 UTC m=+752.455083155" observedRunningTime="2026-01-22 11:52:59.700744888 +0000 UTC m=+753.545815958" watchObservedRunningTime="2026-01-22 11:52:59.70502919 +0000 UTC m=+753.550100260" Jan 22 11:52:59 crc kubenswrapper[4874]: I0122 11:52:59.732087 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8" podStartSLOduration=4.175091831 podStartE2EDuration="19.732068505s" podCreationTimestamp="2026-01-22 11:52:40 +0000 UTC" firstStartedPulling="2026-01-22 11:52:43.023246871 +0000 UTC m=+736.868317941" lastFinishedPulling="2026-01-22 11:52:58.580223545 +0000 UTC m=+752.425294615" observedRunningTime="2026-01-22 11:52:59.728332419 +0000 UTC m=+753.573403489" watchObservedRunningTime="2026-01-22 11:52:59.732068505 +0000 UTC m=+753.577139575" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.479742 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.481110 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.483664 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.483785 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.483898 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.484923 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.485738 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.488659 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-5k545" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.489208 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.489260 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.491173 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.523268 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.617563 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.617627 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.617706 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.617731 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.617777 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.617815 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.617886 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.617911 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.617931 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.617949 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.618048 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.618089 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.618114 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.618133 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.618148 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.718853 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.718914 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.718937 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.718972 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.719003 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.719030 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.719045 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.719061 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.719079 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.719098 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.719115 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.719130 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.719145 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.719169 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.719196 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.720303 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.720443 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.720530 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.720668 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.720755 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.720903 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.721692 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.721799 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.724405 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.724624 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.724662 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.725121 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.725263 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.738059 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.738189 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/c9fa7f48-985a-46b3-a7dc-1b39dfc14243-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"c9fa7f48-985a-46b3-a7dc-1b39dfc14243\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:02 crc kubenswrapper[4874]: I0122 11:53:02.798587 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:03 crc kubenswrapper[4874]: I0122 11:53:03.232462 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.327730 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb"] Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.328530 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.331474 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.331615 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.333128 4874 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-v456v" Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.342440 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb"] Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.458836 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql4qt\" (UniqueName: \"kubernetes.io/projected/e37c03ae-9bcf-4538-9aba-072106273e29-kube-api-access-ql4qt\") pod \"cert-manager-operator-controller-manager-5446d6888b-zhfmb\" (UID: \"e37c03ae-9bcf-4538-9aba-072106273e29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.459248 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e37c03ae-9bcf-4538-9aba-072106273e29-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-zhfmb\" (UID: \"e37c03ae-9bcf-4538-9aba-072106273e29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.560189 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e37c03ae-9bcf-4538-9aba-072106273e29-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-zhfmb\" (UID: \"e37c03ae-9bcf-4538-9aba-072106273e29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.560285 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql4qt\" (UniqueName: \"kubernetes.io/projected/e37c03ae-9bcf-4538-9aba-072106273e29-kube-api-access-ql4qt\") pod \"cert-manager-operator-controller-manager-5446d6888b-zhfmb\" (UID: \"e37c03ae-9bcf-4538-9aba-072106273e29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.561007 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e37c03ae-9bcf-4538-9aba-072106273e29-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-zhfmb\" (UID: \"e37c03ae-9bcf-4538-9aba-072106273e29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.581581 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql4qt\" (UniqueName: \"kubernetes.io/projected/e37c03ae-9bcf-4538-9aba-072106273e29-kube-api-access-ql4qt\") pod \"cert-manager-operator-controller-manager-5446d6888b-zhfmb\" (UID: \"e37c03ae-9bcf-4538-9aba-072106273e29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" Jan 22 11:53:05 crc kubenswrapper[4874]: I0122 11:53:05.647777 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" Jan 22 11:53:07 crc kubenswrapper[4874]: I0122 11:53:07.637105 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c9fa7f48-985a-46b3-a7dc-1b39dfc14243","Type":"ContainerStarted","Data":"bb0384e77d466936e0590d0880f5794b180c8471d296ba55e648b0f5225fc29b"} Jan 22 11:53:07 crc kubenswrapper[4874]: I0122 11:53:07.831404 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb"] Jan 22 11:53:08 crc kubenswrapper[4874]: I0122 11:53:08.642479 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" event={"ID":"e37c03ae-9bcf-4538-9aba-072106273e29","Type":"ContainerStarted","Data":"bc20e22938974036e7e9799cf700a5fa041b9cd37f4b42e9cbc623867237c28f"} Jan 22 11:53:08 crc kubenswrapper[4874]: I0122 11:53:08.643917 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-p9brk" event={"ID":"cd733750-1915-4813-9e44-ec3777ce9c53","Type":"ContainerStarted","Data":"6d63acec90a7b207296bf715fe4ea871ed1bc2eac339709bf80adb3104b5e582"} Jan 22 11:53:08 crc kubenswrapper[4874]: I0122 11:53:08.662326 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-p9brk" podStartSLOduration=9.986926834 podStartE2EDuration="18.662301749s" podCreationTimestamp="2026-01-22 11:52:50 +0000 UTC" firstStartedPulling="2026-01-22 11:52:59.099245859 +0000 UTC m=+752.944316929" lastFinishedPulling="2026-01-22 11:53:07.774620774 +0000 UTC m=+761.619691844" observedRunningTime="2026-01-22 11:53:08.659449091 +0000 UTC m=+762.504520161" watchObservedRunningTime="2026-01-22 11:53:08.662301749 +0000 UTC m=+762.507372819" Jan 22 11:53:11 crc kubenswrapper[4874]: I0122 11:53:11.043759 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-kqrkd" Jan 22 11:53:13 crc kubenswrapper[4874]: I0122 11:53:13.520057 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:53:13 crc kubenswrapper[4874]: I0122 11:53:13.520114 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:53:13 crc kubenswrapper[4874]: I0122 11:53:13.520162 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:53:13 crc kubenswrapper[4874]: I0122 11:53:13.520765 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"58a69e8f9170bdd4dd90e6e773cd03089d2a6279398d2b1a2ba4ed87135be13a"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 11:53:13 crc kubenswrapper[4874]: I0122 11:53:13.520813 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://58a69e8f9170bdd4dd90e6e773cd03089d2a6279398d2b1a2ba4ed87135be13a" gracePeriod=600 Jan 22 11:53:14 crc kubenswrapper[4874]: I0122 11:53:14.684475 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="58a69e8f9170bdd4dd90e6e773cd03089d2a6279398d2b1a2ba4ed87135be13a" exitCode=0 Jan 22 11:53:14 crc kubenswrapper[4874]: I0122 11:53:14.684522 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"58a69e8f9170bdd4dd90e6e773cd03089d2a6279398d2b1a2ba4ed87135be13a"} Jan 22 11:53:14 crc kubenswrapper[4874]: I0122 11:53:14.684568 4874 scope.go:117] "RemoveContainer" containerID="4ef8bf2a0eb4af528f4de7cda955aa92b0de433bdf6d0ebda36103e1834fe2b2" Jan 22 11:53:21 crc kubenswrapper[4874]: E0122 11:53:21.004425 4874 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:fa8de363ab4435c1085ac37f1bad488828c6ae8ba361c5f865c27ef577610911" Jan 22 11:53:21 crc kubenswrapper[4874]: E0122 11:53:21.005041 4874 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-operator,Image:registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:fa8de363ab4435c1085ac37f1bad488828c6ae8ba361c5f865c27ef577610911,Command:[/usr/bin/cert-manager-operator],Args:[start --v=$(OPERATOR_LOG_LEVEL) --trusted-ca-configmap=$(TRUSTED_CA_CONFIGMAP_NAME) --cloud-credentials-secret=$(CLOUD_CREDENTIALS_SECRET_NAME) --unsupported-addon-features=$(UNSUPPORTED_ADDON_FEATURES)],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:cert-manager-operator,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_WEBHOOK,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CA_INJECTOR,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CONTROLLER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ACMESOLVER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:ba937fc4b9eee31422914352c11a45b90754ba4fbe490ea45249b90afdc4e0a7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ISTIOCSR,Value:registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:af1ac813b8ee414ef215936f05197bc498bccbd540f3e2a93cb522221ba112bc,ValueFrom:nil,},EnvVar{Name:OPERAND_IMAGE_VERSION,Value:1.18.3,ValueFrom:nil,},EnvVar{Name:ISTIOCSR_OPERAND_IMAGE_VERSION,Value:0.14.2,ValueFrom:nil,},EnvVar{Name:OPERATOR_IMAGE_VERSION,Value:1.18.0,ValueFrom:nil,},EnvVar{Name:OPERATOR_LOG_LEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:TRUSTED_CA_CONFIGMAP_NAME,Value:,ValueFrom:nil,},EnvVar{Name:CLOUD_CREDENTIALS_SECRET_NAME,Value:,ValueFrom:nil,},EnvVar{Name:UNSUPPORTED_ADDON_FEATURES,Value:,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cert-manager-operator.v1.18.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{33554432 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tmp,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ql4qt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000680000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-operator-controller-manager-5446d6888b-zhfmb_cert-manager-operator(e37c03ae-9bcf-4538-9aba-072106273e29): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 11:53:21 crc kubenswrapper[4874]: E0122 11:53:21.006992 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" podUID="e37c03ae-9bcf-4538-9aba-072106273e29" Jan 22 11:53:21 crc kubenswrapper[4874]: E0122 11:53:21.723114 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:fa8de363ab4435c1085ac37f1bad488828c6ae8ba361c5f865c27ef577610911\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" podUID="e37c03ae-9bcf-4538-9aba-072106273e29" Jan 22 11:53:25 crc kubenswrapper[4874]: I0122 11:53:25.746447 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"2554a3567b106d7ded370e89a12bf13dafb3f02d930bcb0aa478a1f4bf2cf32b"} Jan 22 11:53:25 crc kubenswrapper[4874]: I0122 11:53:25.748486 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c9fa7f48-985a-46b3-a7dc-1b39dfc14243","Type":"ContainerStarted","Data":"76f1f52f33f85f0ce7add7dbcada81bb14335fa3c949ab0293c2a94b1081484e"} Jan 22 11:53:26 crc kubenswrapper[4874]: I0122 11:53:26.123163 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 22 11:53:26 crc kubenswrapper[4874]: I0122 11:53:26.171294 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 22 11:53:27 crc kubenswrapper[4874]: I0122 11:53:27.761348 4874 generic.go:334] "Generic (PLEG): container finished" podID="c9fa7f48-985a-46b3-a7dc-1b39dfc14243" containerID="76f1f52f33f85f0ce7add7dbcada81bb14335fa3c949ab0293c2a94b1081484e" exitCode=0 Jan 22 11:53:27 crc kubenswrapper[4874]: I0122 11:53:27.761452 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c9fa7f48-985a-46b3-a7dc-1b39dfc14243","Type":"ContainerDied","Data":"76f1f52f33f85f0ce7add7dbcada81bb14335fa3c949ab0293c2a94b1081484e"} Jan 22 11:53:28 crc kubenswrapper[4874]: I0122 11:53:28.770265 4874 generic.go:334] "Generic (PLEG): container finished" podID="c9fa7f48-985a-46b3-a7dc-1b39dfc14243" containerID="758db578e91bb6515d0e82c635722c1453210d78166d82b004ff9c8910119fbb" exitCode=0 Jan 22 11:53:28 crc kubenswrapper[4874]: I0122 11:53:28.770308 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c9fa7f48-985a-46b3-a7dc-1b39dfc14243","Type":"ContainerDied","Data":"758db578e91bb6515d0e82c635722c1453210d78166d82b004ff9c8910119fbb"} Jan 22 11:53:29 crc kubenswrapper[4874]: I0122 11:53:29.779518 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c9fa7f48-985a-46b3-a7dc-1b39dfc14243","Type":"ContainerStarted","Data":"b4903bb13070f48c4d782896c6a57a6a3155d42b060fa318be80e908fb0f39c3"} Jan 22 11:53:29 crc kubenswrapper[4874]: I0122 11:53:29.780095 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:29 crc kubenswrapper[4874]: I0122 11:53:29.843827 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=10.705030057 podStartE2EDuration="27.843810046s" podCreationTimestamp="2026-01-22 11:53:02 +0000 UTC" firstStartedPulling="2026-01-22 11:53:07.607980639 +0000 UTC m=+761.453051709" lastFinishedPulling="2026-01-22 11:53:24.746760628 +0000 UTC m=+778.591831698" observedRunningTime="2026-01-22 11:53:29.842827235 +0000 UTC m=+783.687898345" watchObservedRunningTime="2026-01-22 11:53:29.843810046 +0000 UTC m=+783.688881116" Jan 22 11:53:35 crc kubenswrapper[4874]: I0122 11:53:35.825009 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" event={"ID":"e37c03ae-9bcf-4538-9aba-072106273e29","Type":"ContainerStarted","Data":"f0d0c6331a5fb1c370792881f12b84b77bda275fa555f2d0fc6a5cc4fe381538"} Jan 22 11:53:35 crc kubenswrapper[4874]: I0122 11:53:35.853173 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-zhfmb" podStartSLOduration=3.243297462 podStartE2EDuration="30.853153351s" podCreationTimestamp="2026-01-22 11:53:05 +0000 UTC" firstStartedPulling="2026-01-22 11:53:07.83798973 +0000 UTC m=+761.683060800" lastFinishedPulling="2026-01-22 11:53:35.447845609 +0000 UTC m=+789.292916689" observedRunningTime="2026-01-22 11:53:35.852467271 +0000 UTC m=+789.697538351" watchObservedRunningTime="2026-01-22 11:53:35.853153351 +0000 UTC m=+789.698224421" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.863309 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.881518 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.881684 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.884111 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-sys-config" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.884281 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-ca" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.884614 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-jmkbp" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.884643 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-global-ca" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.918586 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.918641 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.918671 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.918868 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-push\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.918986 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.919055 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.919088 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.919125 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.919151 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.919170 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.919204 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcbzf\" (UniqueName: \"kubernetes.io/projected/1b09a4fd-742d-4043-9068-e0c2cfb31f33-kube-api-access-rcbzf\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:38 crc kubenswrapper[4874]: I0122 11:53:38.919247 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020556 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-push\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020625 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020667 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020695 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020736 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020761 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020784 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020811 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcbzf\" (UniqueName: \"kubernetes.io/projected/1b09a4fd-742d-4043-9068-e0c2cfb31f33-kube-api-access-rcbzf\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020840 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020869 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020905 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.020938 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.022032 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.022533 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.022820 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.022990 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.023017 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.023157 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.028228 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.028881 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.029108 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.029800 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.033226 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-push\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.049292 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcbzf\" (UniqueName: \"kubernetes.io/projected/1b09a4fd-742d-4043-9068-e0c2cfb31f33-kube-api-access-rcbzf\") pod \"service-telemetry-operator-1-build\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.200994 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.670601 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 11:53:39 crc kubenswrapper[4874]: I0122 11:53:39.854736 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"1b09a4fd-742d-4043-9068-e0c2cfb31f33","Type":"ContainerStarted","Data":"10c98d3b8b81f049d9113119ec9418b2eec58e5f22be050a8e70df9f0b424d91"} Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.315796 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-cvlzs"] Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.316503 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.318728 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.318861 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.321629 4874 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-2cbsm" Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.350510 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-cvlzs"] Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.444550 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bef508e9-b648-4c62-bc0d-91bf604067ed-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-cvlzs\" (UID: \"bef508e9-b648-4c62-bc0d-91bf604067ed\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.444595 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxkpx\" (UniqueName: \"kubernetes.io/projected/bef508e9-b648-4c62-bc0d-91bf604067ed-kube-api-access-lxkpx\") pod \"cert-manager-webhook-f4fb5df64-cvlzs\" (UID: \"bef508e9-b648-4c62-bc0d-91bf604067ed\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.545962 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bef508e9-b648-4c62-bc0d-91bf604067ed-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-cvlzs\" (UID: \"bef508e9-b648-4c62-bc0d-91bf604067ed\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.546009 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxkpx\" (UniqueName: \"kubernetes.io/projected/bef508e9-b648-4c62-bc0d-91bf604067ed-kube-api-access-lxkpx\") pod \"cert-manager-webhook-f4fb5df64-cvlzs\" (UID: \"bef508e9-b648-4c62-bc0d-91bf604067ed\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.575384 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxkpx\" (UniqueName: \"kubernetes.io/projected/bef508e9-b648-4c62-bc0d-91bf604067ed-kube-api-access-lxkpx\") pod \"cert-manager-webhook-f4fb5df64-cvlzs\" (UID: \"bef508e9-b648-4c62-bc0d-91bf604067ed\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.588473 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bef508e9-b648-4c62-bc0d-91bf604067ed-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-cvlzs\" (UID: \"bef508e9-b648-4c62-bc0d-91bf604067ed\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.633275 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.818300 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-cvlzs"] Jan 22 11:53:40 crc kubenswrapper[4874]: W0122 11:53:40.823886 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbef508e9_b648_4c62_bc0d_91bf604067ed.slice/crio-cc71356a99702a6207e9fd2dc646169e87ad072ef05c8409f7680ebd1f6455e4 WatchSource:0}: Error finding container cc71356a99702a6207e9fd2dc646169e87ad072ef05c8409f7680ebd1f6455e4: Status 404 returned error can't find the container with id cc71356a99702a6207e9fd2dc646169e87ad072ef05c8409f7680ebd1f6455e4 Jan 22 11:53:40 crc kubenswrapper[4874]: I0122 11:53:40.861195 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" event={"ID":"bef508e9-b648-4c62-bc0d-91bf604067ed","Type":"ContainerStarted","Data":"cc71356a99702a6207e9fd2dc646169e87ad072ef05c8409f7680ebd1f6455e4"} Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.123297 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m"] Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.146358 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m" Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.148883 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m"] Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.151676 4874 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-b5dq9" Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.182808 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfmh7\" (UniqueName: \"kubernetes.io/projected/b35625a1-4b85-4915-8ac0-97ff11979513-kube-api-access-mfmh7\") pod \"cert-manager-cainjector-855d9ccff4-mnt2m\" (UID: \"b35625a1-4b85-4915-8ac0-97ff11979513\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m" Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.182883 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b35625a1-4b85-4915-8ac0-97ff11979513-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-mnt2m\" (UID: \"b35625a1-4b85-4915-8ac0-97ff11979513\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m" Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.283657 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b35625a1-4b85-4915-8ac0-97ff11979513-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-mnt2m\" (UID: \"b35625a1-4b85-4915-8ac0-97ff11979513\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m" Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.283820 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfmh7\" (UniqueName: \"kubernetes.io/projected/b35625a1-4b85-4915-8ac0-97ff11979513-kube-api-access-mfmh7\") pod \"cert-manager-cainjector-855d9ccff4-mnt2m\" (UID: \"b35625a1-4b85-4915-8ac0-97ff11979513\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m" Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.316603 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b35625a1-4b85-4915-8ac0-97ff11979513-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-mnt2m\" (UID: \"b35625a1-4b85-4915-8ac0-97ff11979513\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m" Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.316747 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfmh7\" (UniqueName: \"kubernetes.io/projected/b35625a1-4b85-4915-8ac0-97ff11979513-kube-api-access-mfmh7\") pod \"cert-manager-cainjector-855d9ccff4-mnt2m\" (UID: \"b35625a1-4b85-4915-8ac0-97ff11979513\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m" Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.472104 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m" Jan 22 11:53:43 crc kubenswrapper[4874]: I0122 11:53:43.503068 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Jan 22 11:53:49 crc kubenswrapper[4874]: I0122 11:53:49.290726 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 11:53:50 crc kubenswrapper[4874]: I0122 11:53:50.666793 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-7nhkf"] Jan 22 11:53:50 crc kubenswrapper[4874]: I0122 11:53:50.667765 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-7nhkf" Jan 22 11:53:50 crc kubenswrapper[4874]: I0122 11:53:50.673706 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-7nhkf"] Jan 22 11:53:50 crc kubenswrapper[4874]: I0122 11:53:50.674296 4874 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-496mh" Jan 22 11:53:50 crc kubenswrapper[4874]: I0122 11:53:50.700001 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bl67\" (UniqueName: \"kubernetes.io/projected/50438a7c-3a86-4aeb-a9c6-10ccb68f4593-kube-api-access-7bl67\") pod \"cert-manager-86cb77c54b-7nhkf\" (UID: \"50438a7c-3a86-4aeb-a9c6-10ccb68f4593\") " pod="cert-manager/cert-manager-86cb77c54b-7nhkf" Jan 22 11:53:50 crc kubenswrapper[4874]: I0122 11:53:50.700067 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/50438a7c-3a86-4aeb-a9c6-10ccb68f4593-bound-sa-token\") pod \"cert-manager-86cb77c54b-7nhkf\" (UID: \"50438a7c-3a86-4aeb-a9c6-10ccb68f4593\") " pod="cert-manager/cert-manager-86cb77c54b-7nhkf" Jan 22 11:53:50 crc kubenswrapper[4874]: I0122 11:53:50.802188 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bl67\" (UniqueName: \"kubernetes.io/projected/50438a7c-3a86-4aeb-a9c6-10ccb68f4593-kube-api-access-7bl67\") pod \"cert-manager-86cb77c54b-7nhkf\" (UID: \"50438a7c-3a86-4aeb-a9c6-10ccb68f4593\") " pod="cert-manager/cert-manager-86cb77c54b-7nhkf" Jan 22 11:53:50 crc kubenswrapper[4874]: I0122 11:53:50.802303 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/50438a7c-3a86-4aeb-a9c6-10ccb68f4593-bound-sa-token\") pod \"cert-manager-86cb77c54b-7nhkf\" (UID: \"50438a7c-3a86-4aeb-a9c6-10ccb68f4593\") " pod="cert-manager/cert-manager-86cb77c54b-7nhkf" Jan 22 11:53:50 crc kubenswrapper[4874]: I0122 11:53:50.833366 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/50438a7c-3a86-4aeb-a9c6-10ccb68f4593-bound-sa-token\") pod \"cert-manager-86cb77c54b-7nhkf\" (UID: \"50438a7c-3a86-4aeb-a9c6-10ccb68f4593\") " pod="cert-manager/cert-manager-86cb77c54b-7nhkf" Jan 22 11:53:50 crc kubenswrapper[4874]: I0122 11:53:50.834982 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bl67\" (UniqueName: \"kubernetes.io/projected/50438a7c-3a86-4aeb-a9c6-10ccb68f4593-kube-api-access-7bl67\") pod \"cert-manager-86cb77c54b-7nhkf\" (UID: \"50438a7c-3a86-4aeb-a9c6-10ccb68f4593\") " pod="cert-manager/cert-manager-86cb77c54b-7nhkf" Jan 22 11:53:50 crc kubenswrapper[4874]: I0122 11:53:50.983691 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-7nhkf" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.202775 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m"] Jan 22 11:53:51 crc kubenswrapper[4874]: W0122 11:53:51.207386 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb35625a1_4b85_4915_8ac0_97ff11979513.slice/crio-4a563be5ef7eaf5bb94f189fa58db89c1e68323cebfaad76e8bd9efd70ad74cc WatchSource:0}: Error finding container 4a563be5ef7eaf5bb94f189fa58db89c1e68323cebfaad76e8bd9efd70ad74cc: Status 404 returned error can't find the container with id 4a563be5ef7eaf5bb94f189fa58db89c1e68323cebfaad76e8bd9efd70ad74cc Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.388386 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.404804 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.410994 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-global-ca" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.411195 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-sys-config" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.411318 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-ca" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.420780 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.496990 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-7nhkf"] Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513537 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513598 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513633 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-push\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513661 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513683 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513721 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb5k6\" (UniqueName: \"kubernetes.io/projected/b6347418-07a3-41af-aea9-1eddb77e64fb-kube-api-access-kb5k6\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513750 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513775 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513809 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513832 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513857 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.513899 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615096 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615140 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-push\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615158 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615174 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615204 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb5k6\" (UniqueName: \"kubernetes.io/projected/b6347418-07a3-41af-aea9-1eddb77e64fb-kube-api-access-kb5k6\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615226 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615246 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615269 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615289 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615306 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615337 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615367 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615695 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.615744 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.616013 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.616350 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.616577 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.616899 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.616898 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.617074 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.617179 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.620920 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.620983 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-push\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.633860 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb5k6\" (UniqueName: \"kubernetes.io/projected/b6347418-07a3-41af-aea9-1eddb77e64fb-kube-api-access-kb5k6\") pod \"service-telemetry-operator-2-build\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.731083 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.953197 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"1b09a4fd-742d-4043-9068-e0c2cfb31f33","Type":"ContainerStarted","Data":"11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff"} Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.953315 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="1b09a4fd-742d-4043-9068-e0c2cfb31f33" containerName="manage-dockerfile" containerID="cri-o://11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff" gracePeriod=30 Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.954978 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m" event={"ID":"b35625a1-4b85-4915-8ac0-97ff11979513","Type":"ContainerStarted","Data":"c670552933f7270daf00e6564c7c223dbc80cdba112b0bbbdbfc9bdbd5471310"} Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.955016 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m" event={"ID":"b35625a1-4b85-4915-8ac0-97ff11979513","Type":"ContainerStarted","Data":"4a563be5ef7eaf5bb94f189fa58db89c1e68323cebfaad76e8bd9efd70ad74cc"} Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.959115 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" event={"ID":"bef508e9-b648-4c62-bc0d-91bf604067ed","Type":"ContainerStarted","Data":"5ebbd4a950f408f1e40965d49fcc41103d96f3bbd2f95124a8b83dd5a1611bcc"} Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.959276 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.960412 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-7nhkf" event={"ID":"50438a7c-3a86-4aeb-a9c6-10ccb68f4593","Type":"ContainerStarted","Data":"e1c4b6578f52fc6594e2e874b18303c7e8351947f777186a48184d682c938010"} Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.960448 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-7nhkf" event={"ID":"50438a7c-3a86-4aeb-a9c6-10ccb68f4593","Type":"ContainerStarted","Data":"963fe171f462cde36ba653026c68d7571de2986c471df1abe768561e21a30d8a"} Jan 22 11:53:51 crc kubenswrapper[4874]: I0122 11:53:51.974899 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.033715 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-7nhkf" podStartSLOduration=2.033694993 podStartE2EDuration="2.033694993s" podCreationTimestamp="2026-01-22 11:53:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:53:52.00868727 +0000 UTC m=+805.853758340" watchObservedRunningTime="2026-01-22 11:53:52.033694993 +0000 UTC m=+805.878766063" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.035284 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" podStartSLOduration=1.8041309829999999 podStartE2EDuration="12.035275622s" podCreationTimestamp="2026-01-22 11:53:40 +0000 UTC" firstStartedPulling="2026-01-22 11:53:40.826090621 +0000 UTC m=+794.671161691" lastFinishedPulling="2026-01-22 11:53:51.05723526 +0000 UTC m=+804.902306330" observedRunningTime="2026-01-22 11:53:52.029621987 +0000 UTC m=+805.874693067" watchObservedRunningTime="2026-01-22 11:53:52.035275622 +0000 UTC m=+805.880346692" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.046568 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-mnt2m" podStartSLOduration=9.04655092 podStartE2EDuration="9.04655092s" podCreationTimestamp="2026-01-22 11:53:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:53:52.044996583 +0000 UTC m=+805.890067663" watchObservedRunningTime="2026-01-22 11:53:52.04655092 +0000 UTC m=+805.891621990" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.376608 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_1b09a4fd-742d-4043-9068-e0c2cfb31f33/manage-dockerfile/0.log" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.376667 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434124 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildworkdir\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434165 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-blob-cache\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434211 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-system-configs\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434233 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-node-pullsecrets\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434261 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-root\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434282 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcbzf\" (UniqueName: \"kubernetes.io/projected/1b09a4fd-742d-4043-9068-e0c2cfb31f33-kube-api-access-rcbzf\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434303 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-push\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434327 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-ca-bundles\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434377 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-proxy-ca-bundles\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434421 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildcachedir\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434441 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-run\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.434459 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-pull\") pod \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\" (UID: \"1b09a4fd-742d-4043-9068-e0c2cfb31f33\") " Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.435895 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.435940 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.435950 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.435953 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.435736 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.436127 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.436115 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.436241 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.436521 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.440081 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-pull" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-pull") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "builder-dockercfg-jmkbp-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.440206 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b09a4fd-742d-4043-9068-e0c2cfb31f33-kube-api-access-rcbzf" (OuterVolumeSpecName: "kube-api-access-rcbzf") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "kube-api-access-rcbzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.440426 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-push" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-push") pod "1b09a4fd-742d-4043-9068-e0c2cfb31f33" (UID: "1b09a4fd-742d-4043-9068-e0c2cfb31f33"). InnerVolumeSpecName "builder-dockercfg-jmkbp-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535575 4874 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535617 4874 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535633 4874 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535645 4874 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535657 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535668 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcbzf\" (UniqueName: \"kubernetes.io/projected/1b09a4fd-742d-4043-9068-e0c2cfb31f33-kube-api-access-rcbzf\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535678 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-push\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535690 4874 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535700 4874 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b09a4fd-742d-4043-9068-e0c2cfb31f33-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535710 4874 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/1b09a4fd-742d-4043-9068-e0c2cfb31f33-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535720 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/1b09a4fd-742d-4043-9068-e0c2cfb31f33-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.535730 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/1b09a4fd-742d-4043-9068-e0c2cfb31f33-builder-dockercfg-jmkbp-pull\") on node \"crc\" DevicePath \"\"" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.966853 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"b6347418-07a3-41af-aea9-1eddb77e64fb","Type":"ContainerStarted","Data":"07d8d2d70b9416f6aeb9ec8e89a4ab652f04c9798b2aba0722577fe2f16d5b0e"} Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.966903 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"b6347418-07a3-41af-aea9-1eddb77e64fb","Type":"ContainerStarted","Data":"bcebca6560efe9e9a2543339a21c60dc4631ab73e752c93dad1062b96446647d"} Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.969646 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_1b09a4fd-742d-4043-9068-e0c2cfb31f33/manage-dockerfile/0.log" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.969699 4874 generic.go:334] "Generic (PLEG): container finished" podID="1b09a4fd-742d-4043-9068-e0c2cfb31f33" containerID="11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff" exitCode=1 Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.969770 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.969773 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"1b09a4fd-742d-4043-9068-e0c2cfb31f33","Type":"ContainerDied","Data":"11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff"} Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.969829 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"1b09a4fd-742d-4043-9068-e0c2cfb31f33","Type":"ContainerDied","Data":"10c98d3b8b81f049d9113119ec9418b2eec58e5f22be050a8e70df9f0b424d91"} Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.969856 4874 scope.go:117] "RemoveContainer" containerID="11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.991281 4874 scope.go:117] "RemoveContainer" containerID="11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff" Jan 22 11:53:52 crc kubenswrapper[4874]: E0122 11:53:52.991729 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff\": container with ID starting with 11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff not found: ID does not exist" containerID="11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff" Jan 22 11:53:52 crc kubenswrapper[4874]: I0122 11:53:52.991764 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff"} err="failed to get container status \"11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff\": rpc error: code = NotFound desc = could not find container \"11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff\": container with ID starting with 11c232a743fd09cd39e768ebb2123d169afbc2c271af710e33ecb1dd42732dff not found: ID does not exist" Jan 22 11:53:53 crc kubenswrapper[4874]: I0122 11:53:53.006943 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 11:53:53 crc kubenswrapper[4874]: I0122 11:53:53.012629 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 11:53:54 crc kubenswrapper[4874]: I0122 11:53:54.727430 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b09a4fd-742d-4043-9068-e0c2cfb31f33" path="/var/lib/kubelet/pods/1b09a4fd-742d-4043-9068-e0c2cfb31f33/volumes" Jan 22 11:54:00 crc kubenswrapper[4874]: I0122 11:54:00.637075 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-cvlzs" Jan 22 11:54:02 crc kubenswrapper[4874]: I0122 11:54:02.028526 4874 generic.go:334] "Generic (PLEG): container finished" podID="b6347418-07a3-41af-aea9-1eddb77e64fb" containerID="07d8d2d70b9416f6aeb9ec8e89a4ab652f04c9798b2aba0722577fe2f16d5b0e" exitCode=0 Jan 22 11:54:02 crc kubenswrapper[4874]: I0122 11:54:02.028632 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"b6347418-07a3-41af-aea9-1eddb77e64fb","Type":"ContainerDied","Data":"07d8d2d70b9416f6aeb9ec8e89a4ab652f04c9798b2aba0722577fe2f16d5b0e"} Jan 22 11:54:03 crc kubenswrapper[4874]: I0122 11:54:03.038351 4874 generic.go:334] "Generic (PLEG): container finished" podID="b6347418-07a3-41af-aea9-1eddb77e64fb" containerID="24440b237869ff1bb9aafcbe4d425a6f81c7f6a2e6e5467fe6067f0986e8b06f" exitCode=0 Jan 22 11:54:03 crc kubenswrapper[4874]: I0122 11:54:03.038452 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"b6347418-07a3-41af-aea9-1eddb77e64fb","Type":"ContainerDied","Data":"24440b237869ff1bb9aafcbe4d425a6f81c7f6a2e6e5467fe6067f0986e8b06f"} Jan 22 11:54:03 crc kubenswrapper[4874]: I0122 11:54:03.085587 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_b6347418-07a3-41af-aea9-1eddb77e64fb/manage-dockerfile/0.log" Jan 22 11:54:04 crc kubenswrapper[4874]: I0122 11:54:04.047100 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"b6347418-07a3-41af-aea9-1eddb77e64fb","Type":"ContainerStarted","Data":"e7358f32b5ffa397582e9f6239f586687e7f578c71f7955b255e87062a74ca0f"} Jan 22 11:54:04 crc kubenswrapper[4874]: I0122 11:54:04.090097 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-2-build" podStartSLOduration=13.09006503 podStartE2EDuration="13.09006503s" podCreationTimestamp="2026-01-22 11:53:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:54:04.078202763 +0000 UTC m=+817.923273893" watchObservedRunningTime="2026-01-22 11:54:04.09006503 +0000 UTC m=+817.935136180" Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.825674 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-98926"] Jan 22 11:54:59 crc kubenswrapper[4874]: E0122 11:54:59.827413 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b09a4fd-742d-4043-9068-e0c2cfb31f33" containerName="manage-dockerfile" Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.827489 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b09a4fd-742d-4043-9068-e0c2cfb31f33" containerName="manage-dockerfile" Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.827661 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b09a4fd-742d-4043-9068-e0c2cfb31f33" containerName="manage-dockerfile" Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.830454 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98926" Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.830918 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-98926"] Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.854944 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52qk5\" (UniqueName: \"kubernetes.io/projected/4c478499-7674-4b38-b0ea-15b5d9bc4702-kube-api-access-52qk5\") pod \"certified-operators-98926\" (UID: \"4c478499-7674-4b38-b0ea-15b5d9bc4702\") " pod="openshift-marketplace/certified-operators-98926" Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.855200 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c478499-7674-4b38-b0ea-15b5d9bc4702-catalog-content\") pod \"certified-operators-98926\" (UID: \"4c478499-7674-4b38-b0ea-15b5d9bc4702\") " pod="openshift-marketplace/certified-operators-98926" Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.855324 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c478499-7674-4b38-b0ea-15b5d9bc4702-utilities\") pod \"certified-operators-98926\" (UID: \"4c478499-7674-4b38-b0ea-15b5d9bc4702\") " pod="openshift-marketplace/certified-operators-98926" Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.956811 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52qk5\" (UniqueName: \"kubernetes.io/projected/4c478499-7674-4b38-b0ea-15b5d9bc4702-kube-api-access-52qk5\") pod \"certified-operators-98926\" (UID: \"4c478499-7674-4b38-b0ea-15b5d9bc4702\") " pod="openshift-marketplace/certified-operators-98926" Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.956877 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c478499-7674-4b38-b0ea-15b5d9bc4702-catalog-content\") pod \"certified-operators-98926\" (UID: \"4c478499-7674-4b38-b0ea-15b5d9bc4702\") " pod="openshift-marketplace/certified-operators-98926" Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.956916 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c478499-7674-4b38-b0ea-15b5d9bc4702-utilities\") pod \"certified-operators-98926\" (UID: \"4c478499-7674-4b38-b0ea-15b5d9bc4702\") " pod="openshift-marketplace/certified-operators-98926" Jan 22 11:54:59 crc kubenswrapper[4874]: I0122 11:54:59.992447 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52qk5\" (UniqueName: \"kubernetes.io/projected/4c478499-7674-4b38-b0ea-15b5d9bc4702-kube-api-access-52qk5\") pod \"certified-operators-98926\" (UID: \"4c478499-7674-4b38-b0ea-15b5d9bc4702\") " pod="openshift-marketplace/certified-operators-98926" Jan 22 11:55:00 crc kubenswrapper[4874]: I0122 11:55:00.323680 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c478499-7674-4b38-b0ea-15b5d9bc4702-utilities\") pod \"certified-operators-98926\" (UID: \"4c478499-7674-4b38-b0ea-15b5d9bc4702\") " pod="openshift-marketplace/certified-operators-98926" Jan 22 11:55:00 crc kubenswrapper[4874]: I0122 11:55:00.323704 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c478499-7674-4b38-b0ea-15b5d9bc4702-catalog-content\") pod \"certified-operators-98926\" (UID: \"4c478499-7674-4b38-b0ea-15b5d9bc4702\") " pod="openshift-marketplace/certified-operators-98926" Jan 22 11:55:00 crc kubenswrapper[4874]: I0122 11:55:00.454974 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98926" Jan 22 11:55:00 crc kubenswrapper[4874]: I0122 11:55:00.656695 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-98926"] Jan 22 11:55:01 crc kubenswrapper[4874]: I0122 11:55:01.476860 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98926" event={"ID":"4c478499-7674-4b38-b0ea-15b5d9bc4702","Type":"ContainerStarted","Data":"e0f3f6d1f5b747eee16a44d88cba427be28f68d393c3fb831a5b58b3a5f51187"} Jan 22 11:55:02 crc kubenswrapper[4874]: I0122 11:55:02.488154 4874 generic.go:334] "Generic (PLEG): container finished" podID="4c478499-7674-4b38-b0ea-15b5d9bc4702" containerID="1f08a9eb23710c352ac2bbcd443d7a323298901463e74e0229f361507efe2996" exitCode=0 Jan 22 11:55:02 crc kubenswrapper[4874]: I0122 11:55:02.488433 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98926" event={"ID":"4c478499-7674-4b38-b0ea-15b5d9bc4702","Type":"ContainerDied","Data":"1f08a9eb23710c352ac2bbcd443d7a323298901463e74e0229f361507efe2996"} Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.782190 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rgf9f"] Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.784366 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.786868 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rgf9f"] Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.807739 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-catalog-content\") pod \"community-operators-rgf9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.807800 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd6s5\" (UniqueName: \"kubernetes.io/projected/4d056385-773b-49d8-b721-7d0162438e9f-kube-api-access-rd6s5\") pod \"community-operators-rgf9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.807855 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-utilities\") pod \"community-operators-rgf9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.908825 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd6s5\" (UniqueName: \"kubernetes.io/projected/4d056385-773b-49d8-b721-7d0162438e9f-kube-api-access-rd6s5\") pod \"community-operators-rgf9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.908955 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-utilities\") pod \"community-operators-rgf9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.909028 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-catalog-content\") pod \"community-operators-rgf9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.909600 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-catalog-content\") pod \"community-operators-rgf9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.910602 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-utilities\") pod \"community-operators-rgf9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:03 crc kubenswrapper[4874]: I0122 11:55:03.929199 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd6s5\" (UniqueName: \"kubernetes.io/projected/4d056385-773b-49d8-b721-7d0162438e9f-kube-api-access-rd6s5\") pod \"community-operators-rgf9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:04 crc kubenswrapper[4874]: I0122 11:55:04.104129 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:04 crc kubenswrapper[4874]: I0122 11:55:04.545750 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rgf9f"] Jan 22 11:55:04 crc kubenswrapper[4874]: W0122 11:55:04.556487 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d056385_773b_49d8_b721_7d0162438e9f.slice/crio-b7a8bc688a5eaec996a6e4d62822b5030c9d56d63849d9f140695f947f81b1f3 WatchSource:0}: Error finding container b7a8bc688a5eaec996a6e4d62822b5030c9d56d63849d9f140695f947f81b1f3: Status 404 returned error can't find the container with id b7a8bc688a5eaec996a6e4d62822b5030c9d56d63849d9f140695f947f81b1f3 Jan 22 11:55:05 crc kubenswrapper[4874]: I0122 11:55:05.505753 4874 generic.go:334] "Generic (PLEG): container finished" podID="4d056385-773b-49d8-b721-7d0162438e9f" containerID="c028ad03fcd0a54943d191f3957825c84b32372e3465fe517f4fcb2b3968f2a2" exitCode=0 Jan 22 11:55:05 crc kubenswrapper[4874]: I0122 11:55:05.505902 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgf9f" event={"ID":"4d056385-773b-49d8-b721-7d0162438e9f","Type":"ContainerDied","Data":"c028ad03fcd0a54943d191f3957825c84b32372e3465fe517f4fcb2b3968f2a2"} Jan 22 11:55:05 crc kubenswrapper[4874]: I0122 11:55:05.506043 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgf9f" event={"ID":"4d056385-773b-49d8-b721-7d0162438e9f","Type":"ContainerStarted","Data":"b7a8bc688a5eaec996a6e4d62822b5030c9d56d63849d9f140695f947f81b1f3"} Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.175491 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mmh7d"] Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.176943 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.193221 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mmh7d"] Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.359736 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-utilities\") pod \"redhat-operators-mmh7d\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.359805 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q74r6\" (UniqueName: \"kubernetes.io/projected/407598a9-9770-4df2-aa40-0b7a2ebb6b74-kube-api-access-q74r6\") pod \"redhat-operators-mmh7d\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.359888 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-catalog-content\") pod \"redhat-operators-mmh7d\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.461115 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-catalog-content\") pod \"redhat-operators-mmh7d\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.461188 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-utilities\") pod \"redhat-operators-mmh7d\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.461217 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q74r6\" (UniqueName: \"kubernetes.io/projected/407598a9-9770-4df2-aa40-0b7a2ebb6b74-kube-api-access-q74r6\") pod \"redhat-operators-mmh7d\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.461760 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-catalog-content\") pod \"redhat-operators-mmh7d\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.461760 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-utilities\") pod \"redhat-operators-mmh7d\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.480827 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q74r6\" (UniqueName: \"kubernetes.io/projected/407598a9-9770-4df2-aa40-0b7a2ebb6b74-kube-api-access-q74r6\") pod \"redhat-operators-mmh7d\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:06 crc kubenswrapper[4874]: I0122 11:55:06.498708 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:07 crc kubenswrapper[4874]: I0122 11:55:07.378075 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mmh7d"] Jan 22 11:55:07 crc kubenswrapper[4874]: I0122 11:55:07.517544 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmh7d" event={"ID":"407598a9-9770-4df2-aa40-0b7a2ebb6b74","Type":"ContainerStarted","Data":"d5440e0f69fa163e2790d5d166c6aabcddecace2bc92b400850bfe5cdb573090"} Jan 22 11:55:07 crc kubenswrapper[4874]: I0122 11:55:07.519918 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98926" event={"ID":"4c478499-7674-4b38-b0ea-15b5d9bc4702","Type":"ContainerStarted","Data":"3423c31fe5f0d3c779afd5ba1f80b7ff941130a3098df2783d7fa827e40e5745"} Jan 22 11:55:08 crc kubenswrapper[4874]: E0122 11:55:08.158775 4874 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d056385_773b_49d8_b721_7d0162438e9f.slice/crio-conmon-61e621791d40f5cf7de14e9a32d33d52133ebb7e9baf8f265b05b722c24ab767.scope\": RecentStats: unable to find data in memory cache]" Jan 22 11:55:08 crc kubenswrapper[4874]: I0122 11:55:08.528005 4874 generic.go:334] "Generic (PLEG): container finished" podID="4c478499-7674-4b38-b0ea-15b5d9bc4702" containerID="3423c31fe5f0d3c779afd5ba1f80b7ff941130a3098df2783d7fa827e40e5745" exitCode=0 Jan 22 11:55:08 crc kubenswrapper[4874]: I0122 11:55:08.528105 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98926" event={"ID":"4c478499-7674-4b38-b0ea-15b5d9bc4702","Type":"ContainerDied","Data":"3423c31fe5f0d3c779afd5ba1f80b7ff941130a3098df2783d7fa827e40e5745"} Jan 22 11:55:08 crc kubenswrapper[4874]: I0122 11:55:08.533285 4874 generic.go:334] "Generic (PLEG): container finished" podID="4d056385-773b-49d8-b721-7d0162438e9f" containerID="61e621791d40f5cf7de14e9a32d33d52133ebb7e9baf8f265b05b722c24ab767" exitCode=0 Jan 22 11:55:08 crc kubenswrapper[4874]: I0122 11:55:08.533430 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgf9f" event={"ID":"4d056385-773b-49d8-b721-7d0162438e9f","Type":"ContainerDied","Data":"61e621791d40f5cf7de14e9a32d33d52133ebb7e9baf8f265b05b722c24ab767"} Jan 22 11:55:08 crc kubenswrapper[4874]: I0122 11:55:08.536168 4874 generic.go:334] "Generic (PLEG): container finished" podID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerID="40c28e2e42abf9261f99365d65326b15740836021a91f4acb26259216ca3e7b1" exitCode=0 Jan 22 11:55:08 crc kubenswrapper[4874]: I0122 11:55:08.536207 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmh7d" event={"ID":"407598a9-9770-4df2-aa40-0b7a2ebb6b74","Type":"ContainerDied","Data":"40c28e2e42abf9261f99365d65326b15740836021a91f4acb26259216ca3e7b1"} Jan 22 11:55:09 crc kubenswrapper[4874]: I0122 11:55:09.543465 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgf9f" event={"ID":"4d056385-773b-49d8-b721-7d0162438e9f","Type":"ContainerStarted","Data":"d491a04b41daea23345a2b813786ba63b9a8a194f6fb063d231c22300029408b"} Jan 22 11:55:09 crc kubenswrapper[4874]: I0122 11:55:09.545758 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmh7d" event={"ID":"407598a9-9770-4df2-aa40-0b7a2ebb6b74","Type":"ContainerStarted","Data":"0bc4b419597059c10171b789a054785cf3bb4929a5925290a6549cdfef194806"} Jan 22 11:55:09 crc kubenswrapper[4874]: I0122 11:55:09.547818 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98926" event={"ID":"4c478499-7674-4b38-b0ea-15b5d9bc4702","Type":"ContainerStarted","Data":"966fb6f463dcf74310bf96d9792bc06193422b6066b0ad13bd2cf39713a58fe7"} Jan 22 11:55:09 crc kubenswrapper[4874]: I0122 11:55:09.576675 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rgf9f" podStartSLOduration=4.738876114 podStartE2EDuration="6.576655681s" podCreationTimestamp="2026-01-22 11:55:03 +0000 UTC" firstStartedPulling="2026-01-22 11:55:07.105561751 +0000 UTC m=+880.950632821" lastFinishedPulling="2026-01-22 11:55:08.943341308 +0000 UTC m=+882.788412388" observedRunningTime="2026-01-22 11:55:09.570118289 +0000 UTC m=+883.415189369" watchObservedRunningTime="2026-01-22 11:55:09.576655681 +0000 UTC m=+883.421726751" Jan 22 11:55:09 crc kubenswrapper[4874]: I0122 11:55:09.612514 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-98926" podStartSLOduration=4.181670771 podStartE2EDuration="10.612494589s" podCreationTimestamp="2026-01-22 11:54:59 +0000 UTC" firstStartedPulling="2026-01-22 11:55:02.490956603 +0000 UTC m=+876.336027673" lastFinishedPulling="2026-01-22 11:55:08.921780421 +0000 UTC m=+882.766851491" observedRunningTime="2026-01-22 11:55:09.589984603 +0000 UTC m=+883.435055673" watchObservedRunningTime="2026-01-22 11:55:09.612494589 +0000 UTC m=+883.457565659" Jan 22 11:55:10 crc kubenswrapper[4874]: I0122 11:55:10.456180 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-98926" Jan 22 11:55:10 crc kubenswrapper[4874]: I0122 11:55:10.456229 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-98926" Jan 22 11:55:10 crc kubenswrapper[4874]: I0122 11:55:10.554278 4874 generic.go:334] "Generic (PLEG): container finished" podID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerID="0bc4b419597059c10171b789a054785cf3bb4929a5925290a6549cdfef194806" exitCode=0 Jan 22 11:55:10 crc kubenswrapper[4874]: I0122 11:55:10.555257 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmh7d" event={"ID":"407598a9-9770-4df2-aa40-0b7a2ebb6b74","Type":"ContainerDied","Data":"0bc4b419597059c10171b789a054785cf3bb4929a5925290a6549cdfef194806"} Jan 22 11:55:11 crc kubenswrapper[4874]: I0122 11:55:11.500915 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-98926" podUID="4c478499-7674-4b38-b0ea-15b5d9bc4702" containerName="registry-server" probeResult="failure" output=< Jan 22 11:55:11 crc kubenswrapper[4874]: timeout: failed to connect service ":50051" within 1s Jan 22 11:55:11 crc kubenswrapper[4874]: > Jan 22 11:55:12 crc kubenswrapper[4874]: I0122 11:55:12.571333 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmh7d" event={"ID":"407598a9-9770-4df2-aa40-0b7a2ebb6b74","Type":"ContainerStarted","Data":"7502d2cf8a786e85152f56fd1e61f57c84773a339414e06e01565cfd0b41ce53"} Jan 22 11:55:12 crc kubenswrapper[4874]: I0122 11:55:12.596831 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mmh7d" podStartSLOduration=4.162854338 podStartE2EDuration="6.596804428s" podCreationTimestamp="2026-01-22 11:55:06 +0000 UTC" firstStartedPulling="2026-01-22 11:55:08.537902152 +0000 UTC m=+882.382973222" lastFinishedPulling="2026-01-22 11:55:10.971852242 +0000 UTC m=+884.816923312" observedRunningTime="2026-01-22 11:55:12.589667346 +0000 UTC m=+886.434738466" watchObservedRunningTime="2026-01-22 11:55:12.596804428 +0000 UTC m=+886.441875508" Jan 22 11:55:14 crc kubenswrapper[4874]: I0122 11:55:14.104361 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:14 crc kubenswrapper[4874]: I0122 11:55:14.104744 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:15 crc kubenswrapper[4874]: I0122 11:55:14.144449 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:15 crc kubenswrapper[4874]: I0122 11:55:14.673690 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:15 crc kubenswrapper[4874]: I0122 11:55:15.570261 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rgf9f"] Jan 22 11:55:16 crc kubenswrapper[4874]: I0122 11:55:16.499772 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:16 crc kubenswrapper[4874]: I0122 11:55:16.499838 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:16 crc kubenswrapper[4874]: I0122 11:55:16.617440 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rgf9f" podUID="4d056385-773b-49d8-b721-7d0162438e9f" containerName="registry-server" containerID="cri-o://d491a04b41daea23345a2b813786ba63b9a8a194f6fb063d231c22300029408b" gracePeriod=2 Jan 22 11:55:17 crc kubenswrapper[4874]: I0122 11:55:17.562337 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mmh7d" podUID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerName="registry-server" probeResult="failure" output=< Jan 22 11:55:17 crc kubenswrapper[4874]: timeout: failed to connect service ":50051" within 1s Jan 22 11:55:17 crc kubenswrapper[4874]: > Jan 22 11:55:20 crc kubenswrapper[4874]: I0122 11:55:20.509359 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-98926" Jan 22 11:55:20 crc kubenswrapper[4874]: I0122 11:55:20.554720 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-98926" Jan 22 11:55:20 crc kubenswrapper[4874]: I0122 11:55:20.690729 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-98926"] Jan 22 11:55:20 crc kubenswrapper[4874]: I0122 11:55:20.741186 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bbdq6"] Jan 22 11:55:20 crc kubenswrapper[4874]: I0122 11:55:20.741472 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bbdq6" podUID="b0a069d7-0600-4490-82a9-2656913f35b7" containerName="registry-server" containerID="cri-o://56ec71b649678b9d06d64da2f108ecc5d7baa2ab0550b12655288c4bbfbcdef7" gracePeriod=2 Jan 22 11:55:22 crc kubenswrapper[4874]: I0122 11:55:22.472101 4874 generic.go:334] "Generic (PLEG): container finished" podID="4d056385-773b-49d8-b721-7d0162438e9f" containerID="d491a04b41daea23345a2b813786ba63b9a8a194f6fb063d231c22300029408b" exitCode=0 Jan 22 11:55:22 crc kubenswrapper[4874]: I0122 11:55:22.472198 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgf9f" event={"ID":"4d056385-773b-49d8-b721-7d0162438e9f","Type":"ContainerDied","Data":"d491a04b41daea23345a2b813786ba63b9a8a194f6fb063d231c22300029408b"} Jan 22 11:55:22 crc kubenswrapper[4874]: I0122 11:55:22.950033 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:23 crc kubenswrapper[4874]: I0122 11:55:23.090253 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-utilities\") pod \"4d056385-773b-49d8-b721-7d0162438e9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " Jan 22 11:55:23 crc kubenswrapper[4874]: I0122 11:55:23.090299 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd6s5\" (UniqueName: \"kubernetes.io/projected/4d056385-773b-49d8-b721-7d0162438e9f-kube-api-access-rd6s5\") pod \"4d056385-773b-49d8-b721-7d0162438e9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " Jan 22 11:55:23 crc kubenswrapper[4874]: I0122 11:55:23.090411 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-catalog-content\") pod \"4d056385-773b-49d8-b721-7d0162438e9f\" (UID: \"4d056385-773b-49d8-b721-7d0162438e9f\") " Jan 22 11:55:23 crc kubenswrapper[4874]: I0122 11:55:23.091179 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-utilities" (OuterVolumeSpecName: "utilities") pod "4d056385-773b-49d8-b721-7d0162438e9f" (UID: "4d056385-773b-49d8-b721-7d0162438e9f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:55:23 crc kubenswrapper[4874]: I0122 11:55:23.104634 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d056385-773b-49d8-b721-7d0162438e9f-kube-api-access-rd6s5" (OuterVolumeSpecName: "kube-api-access-rd6s5") pod "4d056385-773b-49d8-b721-7d0162438e9f" (UID: "4d056385-773b-49d8-b721-7d0162438e9f"). InnerVolumeSpecName "kube-api-access-rd6s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:55:23 crc kubenswrapper[4874]: I0122 11:55:23.192340 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:23 crc kubenswrapper[4874]: I0122 11:55:23.192385 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd6s5\" (UniqueName: \"kubernetes.io/projected/4d056385-773b-49d8-b721-7d0162438e9f-kube-api-access-rd6s5\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:23 crc kubenswrapper[4874]: I0122 11:55:23.483596 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgf9f" Jan 22 11:55:23 crc kubenswrapper[4874]: I0122 11:55:23.483597 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgf9f" event={"ID":"4d056385-773b-49d8-b721-7d0162438e9f","Type":"ContainerDied","Data":"b7a8bc688a5eaec996a6e4d62822b5030c9d56d63849d9f140695f947f81b1f3"} Jan 22 11:55:23 crc kubenswrapper[4874]: I0122 11:55:23.484078 4874 scope.go:117] "RemoveContainer" containerID="d491a04b41daea23345a2b813786ba63b9a8a194f6fb063d231c22300029408b" Jan 22 11:55:24 crc kubenswrapper[4874]: I0122 11:55:24.111937 4874 scope.go:117] "RemoveContainer" containerID="61e621791d40f5cf7de14e9a32d33d52133ebb7e9baf8f265b05b722c24ab767" Jan 22 11:55:24 crc kubenswrapper[4874]: I0122 11:55:24.134634 4874 scope.go:117] "RemoveContainer" containerID="c028ad03fcd0a54943d191f3957825c84b32372e3465fe517f4fcb2b3968f2a2" Jan 22 11:55:24 crc kubenswrapper[4874]: I0122 11:55:24.381531 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d056385-773b-49d8-b721-7d0162438e9f" (UID: "4d056385-773b-49d8-b721-7d0162438e9f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:55:24 crc kubenswrapper[4874]: I0122 11:55:24.411365 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d056385-773b-49d8-b721-7d0162438e9f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:24 crc kubenswrapper[4874]: I0122 11:55:24.429079 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rgf9f"] Jan 22 11:55:24 crc kubenswrapper[4874]: I0122 11:55:24.434333 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rgf9f"] Jan 22 11:55:24 crc kubenswrapper[4874]: I0122 11:55:24.492122 4874 generic.go:334] "Generic (PLEG): container finished" podID="b0a069d7-0600-4490-82a9-2656913f35b7" containerID="56ec71b649678b9d06d64da2f108ecc5d7baa2ab0550b12655288c4bbfbcdef7" exitCode=0 Jan 22 11:55:24 crc kubenswrapper[4874]: I0122 11:55:24.492167 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbdq6" event={"ID":"b0a069d7-0600-4490-82a9-2656913f35b7","Type":"ContainerDied","Data":"56ec71b649678b9d06d64da2f108ecc5d7baa2ab0550b12655288c4bbfbcdef7"} Jan 22 11:55:24 crc kubenswrapper[4874]: I0122 11:55:24.726541 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d056385-773b-49d8-b721-7d0162438e9f" path="/var/lib/kubelet/pods/4d056385-773b-49d8-b721-7d0162438e9f/volumes" Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.119111 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.221487 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlqxz\" (UniqueName: \"kubernetes.io/projected/b0a069d7-0600-4490-82a9-2656913f35b7-kube-api-access-jlqxz\") pod \"b0a069d7-0600-4490-82a9-2656913f35b7\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.221561 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-utilities\") pod \"b0a069d7-0600-4490-82a9-2656913f35b7\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.221594 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-catalog-content\") pod \"b0a069d7-0600-4490-82a9-2656913f35b7\" (UID: \"b0a069d7-0600-4490-82a9-2656913f35b7\") " Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.222504 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-utilities" (OuterVolumeSpecName: "utilities") pod "b0a069d7-0600-4490-82a9-2656913f35b7" (UID: "b0a069d7-0600-4490-82a9-2656913f35b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.236120 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0a069d7-0600-4490-82a9-2656913f35b7-kube-api-access-jlqxz" (OuterVolumeSpecName: "kube-api-access-jlqxz") pod "b0a069d7-0600-4490-82a9-2656913f35b7" (UID: "b0a069d7-0600-4490-82a9-2656913f35b7"). InnerVolumeSpecName "kube-api-access-jlqxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.277138 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0a069d7-0600-4490-82a9-2656913f35b7" (UID: "b0a069d7-0600-4490-82a9-2656913f35b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.322997 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.323033 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlqxz\" (UniqueName: \"kubernetes.io/projected/b0a069d7-0600-4490-82a9-2656913f35b7-kube-api-access-jlqxz\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.323045 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a069d7-0600-4490-82a9-2656913f35b7-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.498357 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbdq6" event={"ID":"b0a069d7-0600-4490-82a9-2656913f35b7","Type":"ContainerDied","Data":"77b7f983a0caf174b2a1ff9585d546c476923723f895d8d2ccf6e1a80d72b6c1"} Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.498416 4874 scope.go:117] "RemoveContainer" containerID="56ec71b649678b9d06d64da2f108ecc5d7baa2ab0550b12655288c4bbfbcdef7" Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.498471 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bbdq6" Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.515128 4874 scope.go:117] "RemoveContainer" containerID="1ca6b3f9986d9290ed89e908e86e7d2e92aa24a906749df6d2ac85d44a6ba2a1" Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.538474 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bbdq6"] Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.543174 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bbdq6"] Jan 22 11:55:25 crc kubenswrapper[4874]: I0122 11:55:25.543569 4874 scope.go:117] "RemoveContainer" containerID="2136577bcdcbdcffa2ef9ace83cb7011df13fa4d06b362c2ac9b0ca6f8650b12" Jan 22 11:55:26 crc kubenswrapper[4874]: I0122 11:55:26.582275 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:26 crc kubenswrapper[4874]: I0122 11:55:26.627457 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:26 crc kubenswrapper[4874]: I0122 11:55:26.737908 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0a069d7-0600-4490-82a9-2656913f35b7" path="/var/lib/kubelet/pods/b0a069d7-0600-4490-82a9-2656913f35b7/volumes" Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.342017 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mmh7d"] Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.342614 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mmh7d" podUID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerName="registry-server" containerID="cri-o://7502d2cf8a786e85152f56fd1e61f57c84773a339414e06e01565cfd0b41ce53" gracePeriod=2 Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.527326 4874 generic.go:334] "Generic (PLEG): container finished" podID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerID="7502d2cf8a786e85152f56fd1e61f57c84773a339414e06e01565cfd0b41ce53" exitCode=0 Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.527366 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmh7d" event={"ID":"407598a9-9770-4df2-aa40-0b7a2ebb6b74","Type":"ContainerDied","Data":"7502d2cf8a786e85152f56fd1e61f57c84773a339414e06e01565cfd0b41ce53"} Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.705799 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.885749 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-catalog-content\") pod \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.885792 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q74r6\" (UniqueName: \"kubernetes.io/projected/407598a9-9770-4df2-aa40-0b7a2ebb6b74-kube-api-access-q74r6\") pod \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.885883 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-utilities\") pod \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\" (UID: \"407598a9-9770-4df2-aa40-0b7a2ebb6b74\") " Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.886632 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-utilities" (OuterVolumeSpecName: "utilities") pod "407598a9-9770-4df2-aa40-0b7a2ebb6b74" (UID: "407598a9-9770-4df2-aa40-0b7a2ebb6b74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.903076 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/407598a9-9770-4df2-aa40-0b7a2ebb6b74-kube-api-access-q74r6" (OuterVolumeSpecName: "kube-api-access-q74r6") pod "407598a9-9770-4df2-aa40-0b7a2ebb6b74" (UID: "407598a9-9770-4df2-aa40-0b7a2ebb6b74"). InnerVolumeSpecName "kube-api-access-q74r6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.987748 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q74r6\" (UniqueName: \"kubernetes.io/projected/407598a9-9770-4df2-aa40-0b7a2ebb6b74-kube-api-access-q74r6\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.987785 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:29 crc kubenswrapper[4874]: I0122 11:55:29.998278 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "407598a9-9770-4df2-aa40-0b7a2ebb6b74" (UID: "407598a9-9770-4df2-aa40-0b7a2ebb6b74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:55:30 crc kubenswrapper[4874]: I0122 11:55:30.089153 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/407598a9-9770-4df2-aa40-0b7a2ebb6b74-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:30 crc kubenswrapper[4874]: I0122 11:55:30.536072 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmh7d" event={"ID":"407598a9-9770-4df2-aa40-0b7a2ebb6b74","Type":"ContainerDied","Data":"d5440e0f69fa163e2790d5d166c6aabcddecace2bc92b400850bfe5cdb573090"} Jan 22 11:55:30 crc kubenswrapper[4874]: I0122 11:55:30.536126 4874 scope.go:117] "RemoveContainer" containerID="7502d2cf8a786e85152f56fd1e61f57c84773a339414e06e01565cfd0b41ce53" Jan 22 11:55:30 crc kubenswrapper[4874]: I0122 11:55:30.536129 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmh7d" Jan 22 11:55:30 crc kubenswrapper[4874]: I0122 11:55:30.557360 4874 scope.go:117] "RemoveContainer" containerID="0bc4b419597059c10171b789a054785cf3bb4929a5925290a6549cdfef194806" Jan 22 11:55:30 crc kubenswrapper[4874]: I0122 11:55:30.571729 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mmh7d"] Jan 22 11:55:30 crc kubenswrapper[4874]: I0122 11:55:30.571792 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mmh7d"] Jan 22 11:55:30 crc kubenswrapper[4874]: I0122 11:55:30.603374 4874 scope.go:117] "RemoveContainer" containerID="40c28e2e42abf9261f99365d65326b15740836021a91f4acb26259216ca3e7b1" Jan 22 11:55:30 crc kubenswrapper[4874]: I0122 11:55:30.736116 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" path="/var/lib/kubelet/pods/407598a9-9770-4df2-aa40-0b7a2ebb6b74/volumes" Jan 22 11:55:43 crc kubenswrapper[4874]: I0122 11:55:43.521158 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:55:43 crc kubenswrapper[4874]: I0122 11:55:43.521966 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:55:50 crc kubenswrapper[4874]: I0122 11:55:50.686428 4874 generic.go:334] "Generic (PLEG): container finished" podID="b6347418-07a3-41af-aea9-1eddb77e64fb" containerID="e7358f32b5ffa397582e9f6239f586687e7f578c71f7955b255e87062a74ca0f" exitCode=0 Jan 22 11:55:50 crc kubenswrapper[4874]: I0122 11:55:50.687096 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"b6347418-07a3-41af-aea9-1eddb77e64fb","Type":"ContainerDied","Data":"e7358f32b5ffa397582e9f6239f586687e7f578c71f7955b255e87062a74ca0f"} Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.021281 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.186876 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-pull\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.186921 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-buildcachedir\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.186955 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-system-configs\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.186981 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-build-blob-cache\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187031 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-proxy-ca-bundles\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187065 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kb5k6\" (UniqueName: \"kubernetes.io/projected/b6347418-07a3-41af-aea9-1eddb77e64fb-kube-api-access-kb5k6\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187086 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187120 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-run\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187148 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-push\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187169 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-ca-bundles\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187204 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-root\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187261 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-node-pullsecrets\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187300 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-buildworkdir\") pod \"b6347418-07a3-41af-aea9-1eddb77e64fb\" (UID: \"b6347418-07a3-41af-aea9-1eddb77e64fb\") " Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187521 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187769 4874 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.187802 4874 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b6347418-07a3-41af-aea9-1eddb77e64fb-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.188234 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.188298 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.188564 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.188848 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.195688 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6347418-07a3-41af-aea9-1eddb77e64fb-kube-api-access-kb5k6" (OuterVolumeSpecName: "kube-api-access-kb5k6") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "kube-api-access-kb5k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.196169 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-push" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-push") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "builder-dockercfg-jmkbp-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.196442 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-pull" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-pull") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "builder-dockercfg-jmkbp-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.232899 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.288993 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.289037 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-push\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.289055 4874 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.289071 4874 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.289088 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/b6347418-07a3-41af-aea9-1eddb77e64fb-builder-dockercfg-jmkbp-pull\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.289105 4874 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.289121 4874 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6347418-07a3-41af-aea9-1eddb77e64fb-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.289136 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kb5k6\" (UniqueName: \"kubernetes.io/projected/b6347418-07a3-41af-aea9-1eddb77e64fb-kube-api-access-kb5k6\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.401275 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.492659 4874 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.706226 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"b6347418-07a3-41af-aea9-1eddb77e64fb","Type":"ContainerDied","Data":"bcebca6560efe9e9a2543339a21c60dc4631ab73e752c93dad1062b96446647d"} Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.706267 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcebca6560efe9e9a2543339a21c60dc4631ab73e752c93dad1062b96446647d" Jan 22 11:55:52 crc kubenswrapper[4874]: I0122 11:55:52.706342 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 11:55:54 crc kubenswrapper[4874]: I0122 11:55:54.917505 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "b6347418-07a3-41af-aea9-1eddb77e64fb" (UID: "b6347418-07a3-41af-aea9-1eddb77e64fb"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:55:54 crc kubenswrapper[4874]: I0122 11:55:54.936262 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b6347418-07a3-41af-aea9-1eddb77e64fb-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065017 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065613 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerName="extract-content" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065631 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerName="extract-content" Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065647 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d056385-773b-49d8-b721-7d0162438e9f" containerName="registry-server" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065655 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d056385-773b-49d8-b721-7d0162438e9f" containerName="registry-server" Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065667 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d056385-773b-49d8-b721-7d0162438e9f" containerName="extract-utilities" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065676 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d056385-773b-49d8-b721-7d0162438e9f" containerName="extract-utilities" Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065689 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerName="registry-server" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065698 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerName="registry-server" Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065709 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerName="extract-utilities" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065717 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerName="extract-utilities" Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065729 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0a069d7-0600-4490-82a9-2656913f35b7" containerName="extract-content" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065738 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0a069d7-0600-4490-82a9-2656913f35b7" containerName="extract-content" Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065747 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0a069d7-0600-4490-82a9-2656913f35b7" containerName="registry-server" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065756 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0a069d7-0600-4490-82a9-2656913f35b7" containerName="registry-server" Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065768 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6347418-07a3-41af-aea9-1eddb77e64fb" containerName="manage-dockerfile" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065776 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6347418-07a3-41af-aea9-1eddb77e64fb" containerName="manage-dockerfile" Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065786 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6347418-07a3-41af-aea9-1eddb77e64fb" containerName="docker-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065794 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6347418-07a3-41af-aea9-1eddb77e64fb" containerName="docker-build" Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065805 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6347418-07a3-41af-aea9-1eddb77e64fb" containerName="git-clone" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065813 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6347418-07a3-41af-aea9-1eddb77e64fb" containerName="git-clone" Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065826 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d056385-773b-49d8-b721-7d0162438e9f" containerName="extract-content" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065834 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d056385-773b-49d8-b721-7d0162438e9f" containerName="extract-content" Jan 22 11:55:57 crc kubenswrapper[4874]: E0122 11:55:57.065847 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0a069d7-0600-4490-82a9-2656913f35b7" containerName="extract-utilities" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065855 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0a069d7-0600-4490-82a9-2656913f35b7" containerName="extract-utilities" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065978 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d056385-773b-49d8-b721-7d0162438e9f" containerName="registry-server" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.065991 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6347418-07a3-41af-aea9-1eddb77e64fb" containerName="docker-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.066008 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="407598a9-9770-4df2-aa40-0b7a2ebb6b74" containerName="registry-server" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.066019 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0a069d7-0600-4490-82a9-2656913f35b7" containerName="registry-server" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.066812 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.069266 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-global-ca" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.069529 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-jmkbp" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.069558 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-sys-config" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.070421 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-ca" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.079464 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.079504 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.079522 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sjlq\" (UniqueName: \"kubernetes.io/projected/25c1a578-3994-43bb-abfb-5dcfbac56652-kube-api-access-5sjlq\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.079550 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.079642 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.079684 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.079719 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-push\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.079752 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.079776 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.079904 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.079984 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.080019 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.081843 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180351 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180419 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180462 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180481 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sjlq\" (UniqueName: \"kubernetes.io/projected/25c1a578-3994-43bb-abfb-5dcfbac56652-kube-api-access-5sjlq\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180499 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180518 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180548 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180574 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180599 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-push\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180644 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180664 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.180707 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.181262 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.181703 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.181883 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.181894 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.181926 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.182229 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.182240 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.182693 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.182939 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.186104 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.189937 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-push\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.199093 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sjlq\" (UniqueName: \"kubernetes.io/projected/25c1a578-3994-43bb-abfb-5dcfbac56652-kube-api-access-5sjlq\") pod \"smart-gateway-operator-1-build\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.390968 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:55:57 crc kubenswrapper[4874]: I0122 11:55:57.876514 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 22 11:55:58 crc kubenswrapper[4874]: I0122 11:55:58.779895 4874 generic.go:334] "Generic (PLEG): container finished" podID="25c1a578-3994-43bb-abfb-5dcfbac56652" containerID="e26549d52ce98cf01bc9efdf9791b993b46a6268920f6328a10f303f52ec5295" exitCode=0 Jan 22 11:55:58 crc kubenswrapper[4874]: I0122 11:55:58.779953 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"25c1a578-3994-43bb-abfb-5dcfbac56652","Type":"ContainerDied","Data":"e26549d52ce98cf01bc9efdf9791b993b46a6268920f6328a10f303f52ec5295"} Jan 22 11:55:58 crc kubenswrapper[4874]: I0122 11:55:58.780006 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"25c1a578-3994-43bb-abfb-5dcfbac56652","Type":"ContainerStarted","Data":"5133f382caa3ada8209b89766e9dbab4bde7b156a4c6c26bc6d027ea0f1922b9"} Jan 22 11:55:59 crc kubenswrapper[4874]: I0122 11:55:59.794585 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"25c1a578-3994-43bb-abfb-5dcfbac56652","Type":"ContainerStarted","Data":"919ff0b784a44bdf0e65017895fbdc7e586d036a33053e72fca76314755ed735"} Jan 22 11:56:08 crc kubenswrapper[4874]: I0122 11:56:08.046334 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-1-build" podStartSLOduration=11.046306004 podStartE2EDuration="11.046306004s" podCreationTimestamp="2026-01-22 11:55:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:55:59.84067771 +0000 UTC m=+933.685748830" watchObservedRunningTime="2026-01-22 11:56:08.046306004 +0000 UTC m=+941.891377114" Jan 22 11:56:08 crc kubenswrapper[4874]: I0122 11:56:08.055745 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 22 11:56:08 crc kubenswrapper[4874]: I0122 11:56:08.056106 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/smart-gateway-operator-1-build" podUID="25c1a578-3994-43bb-abfb-5dcfbac56652" containerName="docker-build" containerID="cri-o://919ff0b784a44bdf0e65017895fbdc7e586d036a33053e72fca76314755ed735" gracePeriod=30 Jan 22 11:56:08 crc kubenswrapper[4874]: I0122 11:56:08.859251 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_25c1a578-3994-43bb-abfb-5dcfbac56652/docker-build/0.log" Jan 22 11:56:08 crc kubenswrapper[4874]: I0122 11:56:08.860460 4874 generic.go:334] "Generic (PLEG): container finished" podID="25c1a578-3994-43bb-abfb-5dcfbac56652" containerID="919ff0b784a44bdf0e65017895fbdc7e586d036a33053e72fca76314755ed735" exitCode=1 Jan 22 11:56:08 crc kubenswrapper[4874]: I0122 11:56:08.860507 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"25c1a578-3994-43bb-abfb-5dcfbac56652","Type":"ContainerDied","Data":"919ff0b784a44bdf0e65017895fbdc7e586d036a33053e72fca76314755ed735"} Jan 22 11:56:08 crc kubenswrapper[4874]: I0122 11:56:08.968860 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_25c1a578-3994-43bb-abfb-5dcfbac56652/docker-build/0.log" Jan 22 11:56:08 crc kubenswrapper[4874]: I0122 11:56:08.969238 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.149511 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-push\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.150550 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-proxy-ca-bundles\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.150584 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-ca-bundles\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.150661 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-root\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.150736 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-pull\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.150798 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-node-pullsecrets\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.150821 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-system-configs\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.150846 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-buildcachedir\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.150878 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-run\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.150916 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-buildworkdir\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.150945 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sjlq\" (UniqueName: \"kubernetes.io/projected/25c1a578-3994-43bb-abfb-5dcfbac56652-kube-api-access-5sjlq\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.150990 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-build-blob-cache\") pod \"25c1a578-3994-43bb-abfb-5dcfbac56652\" (UID: \"25c1a578-3994-43bb-abfb-5dcfbac56652\") " Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.151038 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.151124 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.151569 4874 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.151595 4874 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/25c1a578-3994-43bb-abfb-5dcfbac56652-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.151754 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.151786 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.152672 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.153237 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.153509 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.155932 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-pull" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-pull") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "builder-dockercfg-jmkbp-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.157101 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25c1a578-3994-43bb-abfb-5dcfbac56652-kube-api-access-5sjlq" (OuterVolumeSpecName: "kube-api-access-5sjlq") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "kube-api-access-5sjlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.157778 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-push" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-push") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "builder-dockercfg-jmkbp-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.252900 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-pull\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.252952 4874 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.252973 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.252991 4874 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.253010 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sjlq\" (UniqueName: \"kubernetes.io/projected/25c1a578-3994-43bb-abfb-5dcfbac56652-kube-api-access-5sjlq\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.253028 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/25c1a578-3994-43bb-abfb-5dcfbac56652-builder-dockercfg-jmkbp-push\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.253045 4874 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.253062 4874 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25c1a578-3994-43bb-abfb-5dcfbac56652-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.347729 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.354735 4874 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.659924 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Jan 22 11:56:09 crc kubenswrapper[4874]: E0122 11:56:09.660228 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c1a578-3994-43bb-abfb-5dcfbac56652" containerName="manage-dockerfile" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.660252 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c1a578-3994-43bb-abfb-5dcfbac56652" containerName="manage-dockerfile" Jan 22 11:56:09 crc kubenswrapper[4874]: E0122 11:56:09.660271 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c1a578-3994-43bb-abfb-5dcfbac56652" containerName="docker-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.660281 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c1a578-3994-43bb-abfb-5dcfbac56652" containerName="docker-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.660496 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="25c1a578-3994-43bb-abfb-5dcfbac56652" containerName="docker-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.661886 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.669726 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-global-ca" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.669746 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-sys-config" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.669952 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-ca" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.676343 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.694141 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "25c1a578-3994-43bb-abfb-5dcfbac56652" (UID: "25c1a578-3994-43bb-abfb-5dcfbac56652"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760184 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760305 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760368 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760466 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760498 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760587 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760641 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-push\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760725 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760778 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pz6p\" (UniqueName: \"kubernetes.io/projected/e36b7db7-fd44-4552-8710-adb858c931c9-kube-api-access-2pz6p\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760818 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760851 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760910 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.760999 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/25c1a578-3994-43bb-abfb-5dcfbac56652-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.862526 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.862631 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pz6p\" (UniqueName: \"kubernetes.io/projected/e36b7db7-fd44-4552-8710-adb858c931c9-kube-api-access-2pz6p\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.862689 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.862736 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.862811 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.862892 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.862980 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.863049 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.863095 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.863136 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.863286 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.863357 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-push\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.863573 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.863611 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.863997 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.864574 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.865075 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.865394 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.865926 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.866114 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.866735 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.869496 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.873325 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_25c1a578-3994-43bb-abfb-5dcfbac56652/docker-build/0.log" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.873994 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"25c1a578-3994-43bb-abfb-5dcfbac56652","Type":"ContainerDied","Data":"5133f382caa3ada8209b89766e9dbab4bde7b156a4c6c26bc6d027ea0f1922b9"} Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.874071 4874 scope.go:117] "RemoveContainer" containerID="919ff0b784a44bdf0e65017895fbdc7e586d036a33053e72fca76314755ed735" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.874183 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.883946 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-push\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.888928 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pz6p\" (UniqueName: \"kubernetes.io/projected/e36b7db7-fd44-4552-8710-adb858c931c9-kube-api-access-2pz6p\") pod \"smart-gateway-operator-2-build\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.952924 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.964274 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 22 11:56:09 crc kubenswrapper[4874]: I0122 11:56:09.992185 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:56:10 crc kubenswrapper[4874]: I0122 11:56:10.004721 4874 scope.go:117] "RemoveContainer" containerID="e26549d52ce98cf01bc9efdf9791b993b46a6268920f6328a10f303f52ec5295" Jan 22 11:56:10 crc kubenswrapper[4874]: I0122 11:56:10.502526 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Jan 22 11:56:10 crc kubenswrapper[4874]: W0122 11:56:10.511285 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode36b7db7_fd44_4552_8710_adb858c931c9.slice/crio-a455ec88ef39edfe840d2acca84b9a3be2dc686080708adaf4c9e9cbdfea499a WatchSource:0}: Error finding container a455ec88ef39edfe840d2acca84b9a3be2dc686080708adaf4c9e9cbdfea499a: Status 404 returned error can't find the container with id a455ec88ef39edfe840d2acca84b9a3be2dc686080708adaf4c9e9cbdfea499a Jan 22 11:56:10 crc kubenswrapper[4874]: I0122 11:56:10.726474 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25c1a578-3994-43bb-abfb-5dcfbac56652" path="/var/lib/kubelet/pods/25c1a578-3994-43bb-abfb-5dcfbac56652/volumes" Jan 22 11:56:10 crc kubenswrapper[4874]: I0122 11:56:10.884349 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e36b7db7-fd44-4552-8710-adb858c931c9","Type":"ContainerStarted","Data":"a455ec88ef39edfe840d2acca84b9a3be2dc686080708adaf4c9e9cbdfea499a"} Jan 22 11:56:11 crc kubenswrapper[4874]: I0122 11:56:11.895124 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e36b7db7-fd44-4552-8710-adb858c931c9","Type":"ContainerStarted","Data":"28499c0b98f716958469ca68853b5a713dc17068a8f31f5c05116dd3151a40a9"} Jan 22 11:56:12 crc kubenswrapper[4874]: I0122 11:56:12.903663 4874 generic.go:334] "Generic (PLEG): container finished" podID="e36b7db7-fd44-4552-8710-adb858c931c9" containerID="28499c0b98f716958469ca68853b5a713dc17068a8f31f5c05116dd3151a40a9" exitCode=0 Jan 22 11:56:12 crc kubenswrapper[4874]: I0122 11:56:12.904322 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e36b7db7-fd44-4552-8710-adb858c931c9","Type":"ContainerDied","Data":"28499c0b98f716958469ca68853b5a713dc17068a8f31f5c05116dd3151a40a9"} Jan 22 11:56:13 crc kubenswrapper[4874]: I0122 11:56:13.520684 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:56:13 crc kubenswrapper[4874]: I0122 11:56:13.520742 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:56:13 crc kubenswrapper[4874]: I0122 11:56:13.912922 4874 generic.go:334] "Generic (PLEG): container finished" podID="e36b7db7-fd44-4552-8710-adb858c931c9" containerID="5bc9027fb3ee25faa268acd977fc54b0f5fb85130c768e0e084e4cd0a17cda10" exitCode=0 Jan 22 11:56:13 crc kubenswrapper[4874]: I0122 11:56:13.912960 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e36b7db7-fd44-4552-8710-adb858c931c9","Type":"ContainerDied","Data":"5bc9027fb3ee25faa268acd977fc54b0f5fb85130c768e0e084e4cd0a17cda10"} Jan 22 11:56:13 crc kubenswrapper[4874]: I0122 11:56:13.959875 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-2-build_e36b7db7-fd44-4552-8710-adb858c931c9/manage-dockerfile/0.log" Jan 22 11:56:14 crc kubenswrapper[4874]: I0122 11:56:14.923687 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e36b7db7-fd44-4552-8710-adb858c931c9","Type":"ContainerStarted","Data":"f373e74b504c6b60d88ff77f379cfc029b713c333e26f43f4c2808ee66deffc4"} Jan 22 11:56:14 crc kubenswrapper[4874]: I0122 11:56:14.960314 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-2-build" podStartSLOduration=5.9602894410000005 podStartE2EDuration="5.960289441s" podCreationTimestamp="2026-01-22 11:56:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:56:14.956150102 +0000 UTC m=+948.801221222" watchObservedRunningTime="2026-01-22 11:56:14.960289441 +0000 UTC m=+948.805360531" Jan 22 11:56:43 crc kubenswrapper[4874]: I0122 11:56:43.520252 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:56:43 crc kubenswrapper[4874]: I0122 11:56:43.520838 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:56:43 crc kubenswrapper[4874]: I0122 11:56:43.520892 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 11:56:43 crc kubenswrapper[4874]: I0122 11:56:43.521604 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2554a3567b106d7ded370e89a12bf13dafb3f02d930bcb0aa478a1f4bf2cf32b"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 11:56:43 crc kubenswrapper[4874]: I0122 11:56:43.521675 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://2554a3567b106d7ded370e89a12bf13dafb3f02d930bcb0aa478a1f4bf2cf32b" gracePeriod=600 Jan 22 11:56:44 crc kubenswrapper[4874]: I0122 11:56:44.125008 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="2554a3567b106d7ded370e89a12bf13dafb3f02d930bcb0aa478a1f4bf2cf32b" exitCode=0 Jan 22 11:56:44 crc kubenswrapper[4874]: I0122 11:56:44.125085 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"2554a3567b106d7ded370e89a12bf13dafb3f02d930bcb0aa478a1f4bf2cf32b"} Jan 22 11:56:44 crc kubenswrapper[4874]: I0122 11:56:44.125433 4874 scope.go:117] "RemoveContainer" containerID="58a69e8f9170bdd4dd90e6e773cd03089d2a6279398d2b1a2ba4ed87135be13a" Jan 22 11:56:45 crc kubenswrapper[4874]: I0122 11:56:45.134827 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"8527e378a489c65991df9b62c1d10fce3a020be8e3c0c8d8d62b128fb0466805"} Jan 22 11:57:26 crc kubenswrapper[4874]: I0122 11:57:26.440594 4874 generic.go:334] "Generic (PLEG): container finished" podID="e36b7db7-fd44-4552-8710-adb858c931c9" containerID="f373e74b504c6b60d88ff77f379cfc029b713c333e26f43f4c2808ee66deffc4" exitCode=0 Jan 22 11:57:26 crc kubenswrapper[4874]: I0122 11:57:26.440684 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e36b7db7-fd44-4552-8710-adb858c931c9","Type":"ContainerDied","Data":"f373e74b504c6b60d88ff77f379cfc029b713c333e26f43f4c2808ee66deffc4"} Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.727301 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.927068 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-proxy-ca-bundles\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.927427 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-build-blob-cache\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.927448 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-run\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.927480 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-buildworkdir\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.927498 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pz6p\" (UniqueName: \"kubernetes.io/projected/e36b7db7-fd44-4552-8710-adb858c931c9-kube-api-access-2pz6p\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.928094 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.928484 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.928538 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-node-pullsecrets\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.928598 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.928634 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-system-configs\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.928660 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-ca-bundles\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.928920 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.928978 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-pull\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.929181 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.929278 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-root\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.929285 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.929320 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-buildcachedir\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.929357 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-push\") pod \"e36b7db7-fd44-4552-8710-adb858c931c9\" (UID: \"e36b7db7-fd44-4552-8710-adb858c931c9\") " Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.929741 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.929757 4874 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.929767 4874 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.929776 4874 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.929785 4874 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.929793 4874 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e36b7db7-fd44-4552-8710-adb858c931c9-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.931121 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.951584 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-pull" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-pull") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "builder-dockercfg-jmkbp-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.951640 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-push" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-push") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "builder-dockercfg-jmkbp-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:57:27 crc kubenswrapper[4874]: I0122 11:57:27.966613 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e36b7db7-fd44-4552-8710-adb858c931c9-kube-api-access-2pz6p" (OuterVolumeSpecName: "kube-api-access-2pz6p") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "kube-api-access-2pz6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:57:28 crc kubenswrapper[4874]: I0122 11:57:28.031260 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pz6p\" (UniqueName: \"kubernetes.io/projected/e36b7db7-fd44-4552-8710-adb858c931c9-kube-api-access-2pz6p\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:28 crc kubenswrapper[4874]: I0122 11:57:28.031295 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-pull\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:28 crc kubenswrapper[4874]: I0122 11:57:28.031309 4874 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e36b7db7-fd44-4552-8710-adb858c931c9-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:28 crc kubenswrapper[4874]: I0122 11:57:28.031321 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/e36b7db7-fd44-4552-8710-adb858c931c9-builder-dockercfg-jmkbp-push\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:28 crc kubenswrapper[4874]: I0122 11:57:28.182646 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:57:28 crc kubenswrapper[4874]: I0122 11:57:28.232982 4874 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:28 crc kubenswrapper[4874]: I0122 11:57:28.457521 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e36b7db7-fd44-4552-8710-adb858c931c9","Type":"ContainerDied","Data":"a455ec88ef39edfe840d2acca84b9a3be2dc686080708adaf4c9e9cbdfea499a"} Jan 22 11:57:28 crc kubenswrapper[4874]: I0122 11:57:28.457563 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a455ec88ef39edfe840d2acca84b9a3be2dc686080708adaf4c9e9cbdfea499a" Jan 22 11:57:28 crc kubenswrapper[4874]: I0122 11:57:28.457624 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Jan 22 11:57:30 crc kubenswrapper[4874]: I0122 11:57:30.421952 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "e36b7db7-fd44-4552-8710-adb858c931c9" (UID: "e36b7db7-fd44-4552-8710-adb858c931c9"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:57:30 crc kubenswrapper[4874]: I0122 11:57:30.477110 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e36b7db7-fd44-4552-8710-adb858c931c9-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.179941 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 22 11:57:33 crc kubenswrapper[4874]: E0122 11:57:33.180489 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36b7db7-fd44-4552-8710-adb858c931c9" containerName="manage-dockerfile" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.180505 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36b7db7-fd44-4552-8710-adb858c931c9" containerName="manage-dockerfile" Jan 22 11:57:33 crc kubenswrapper[4874]: E0122 11:57:33.180529 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36b7db7-fd44-4552-8710-adb858c931c9" containerName="git-clone" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.180536 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36b7db7-fd44-4552-8710-adb858c931c9" containerName="git-clone" Jan 22 11:57:33 crc kubenswrapper[4874]: E0122 11:57:33.180549 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36b7db7-fd44-4552-8710-adb858c931c9" containerName="docker-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.180557 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36b7db7-fd44-4552-8710-adb858c931c9" containerName="docker-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.180672 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="e36b7db7-fd44-4552-8710-adb858c931c9" containerName="docker-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.181386 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.184263 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-sys-config" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.184270 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-ca" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.185563 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-global-ca" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.190551 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-jmkbp" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.206284 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.213800 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.213826 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.213864 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.213895 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s26q\" (UniqueName: \"kubernetes.io/projected/fe878840-b584-47fb-aca1-bdbdf8f90b43-kube-api-access-9s26q\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.213959 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildcachedir\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.213975 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.213993 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-push\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.214040 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-pull\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.214062 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-root\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.214085 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-run\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.214115 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildworkdir\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.214138 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-system-configs\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.315510 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.315631 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s26q\" (UniqueName: \"kubernetes.io/projected/fe878840-b584-47fb-aca1-bdbdf8f90b43-kube-api-access-9s26q\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.315704 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildcachedir\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.315756 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.315792 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildcachedir\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.315801 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-push\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.315976 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-pull\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.316655 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-root\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.316721 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-run\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.316749 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildworkdir\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.316794 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-system-configs\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.316863 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.316878 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.317058 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.317143 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-run\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.317068 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.317212 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildworkdir\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.317227 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.317357 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.317572 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-system-configs\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.317649 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-root\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.322624 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-push\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.322909 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-pull\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.343392 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s26q\" (UniqueName: \"kubernetes.io/projected/fe878840-b584-47fb-aca1-bdbdf8f90b43-kube-api-access-9s26q\") pod \"sg-core-1-build\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.502082 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Jan 22 11:57:33 crc kubenswrapper[4874]: I0122 11:57:33.901908 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 22 11:57:34 crc kubenswrapper[4874]: I0122 11:57:34.505032 4874 generic.go:334] "Generic (PLEG): container finished" podID="fe878840-b584-47fb-aca1-bdbdf8f90b43" containerID="89564f4dbf9a2d19321b838663656b209f8663a0ed711a85fb6ba3eae0a0748e" exitCode=0 Jan 22 11:57:34 crc kubenswrapper[4874]: I0122 11:57:34.505118 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"fe878840-b584-47fb-aca1-bdbdf8f90b43","Type":"ContainerDied","Data":"89564f4dbf9a2d19321b838663656b209f8663a0ed711a85fb6ba3eae0a0748e"} Jan 22 11:57:34 crc kubenswrapper[4874]: I0122 11:57:34.505184 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"fe878840-b584-47fb-aca1-bdbdf8f90b43","Type":"ContainerStarted","Data":"24fd5a5e2c73d17a2c7b7ce83a3fc6b3dcc6015ed1e2a7fbe4eb45a26d0d8694"} Jan 22 11:57:35 crc kubenswrapper[4874]: I0122 11:57:35.513603 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"fe878840-b584-47fb-aca1-bdbdf8f90b43","Type":"ContainerStarted","Data":"f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60"} Jan 22 11:57:35 crc kubenswrapper[4874]: I0122 11:57:35.536015 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-1-build" podStartSLOduration=2.535995184 podStartE2EDuration="2.535995184s" podCreationTimestamp="2026-01-22 11:57:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:57:35.532056113 +0000 UTC m=+1029.377127203" watchObservedRunningTime="2026-01-22 11:57:35.535995184 +0000 UTC m=+1029.381066254" Jan 22 11:57:43 crc kubenswrapper[4874]: I0122 11:57:43.605834 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 22 11:57:43 crc kubenswrapper[4874]: I0122 11:57:43.606535 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/sg-core-1-build" podUID="fe878840-b584-47fb-aca1-bdbdf8f90b43" containerName="docker-build" containerID="cri-o://f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60" gracePeriod=30 Jan 22 11:57:43 crc kubenswrapper[4874]: I0122 11:57:43.999668 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_fe878840-b584-47fb-aca1-bdbdf8f90b43/docker-build/0.log" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.000996 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013584 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-proxy-ca-bundles\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013636 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-system-configs\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013683 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-pull\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013723 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-root\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013754 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s26q\" (UniqueName: \"kubernetes.io/projected/fe878840-b584-47fb-aca1-bdbdf8f90b43-kube-api-access-9s26q\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013801 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-node-pullsecrets\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013822 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-push\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013853 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildcachedir\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013881 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-blob-cache\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013936 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-ca-bundles\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013964 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-run\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.013988 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildworkdir\") pod \"fe878840-b584-47fb-aca1-bdbdf8f90b43\" (UID: \"fe878840-b584-47fb-aca1-bdbdf8f90b43\") " Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.014225 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.015173 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.015275 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.015379 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.015380 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.015453 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.016264 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.021290 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-pull" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-pull") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "builder-dockercfg-jmkbp-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.028861 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-push" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-push") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "builder-dockercfg-jmkbp-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.028955 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe878840-b584-47fb-aca1-bdbdf8f90b43-kube-api-access-9s26q" (OuterVolumeSpecName: "kube-api-access-9s26q") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "kube-api-access-9s26q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.115334 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-push\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.115377 4874 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.115389 4874 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.115416 4874 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.115429 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.115440 4874 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.115451 4874 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.115463 4874 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.115474 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/fe878840-b584-47fb-aca1-bdbdf8f90b43-builder-dockercfg-jmkbp-pull\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.115485 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s26q\" (UniqueName: \"kubernetes.io/projected/fe878840-b584-47fb-aca1-bdbdf8f90b43-kube-api-access-9s26q\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.144996 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.216584 4874 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.223632 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "fe878840-b584-47fb-aca1-bdbdf8f90b43" (UID: "fe878840-b584-47fb-aca1-bdbdf8f90b43"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.317885 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/fe878840-b584-47fb-aca1-bdbdf8f90b43-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.578190 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_fe878840-b584-47fb-aca1-bdbdf8f90b43/docker-build/0.log" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.578780 4874 generic.go:334] "Generic (PLEG): container finished" podID="fe878840-b584-47fb-aca1-bdbdf8f90b43" containerID="f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60" exitCode=1 Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.578837 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"fe878840-b584-47fb-aca1-bdbdf8f90b43","Type":"ContainerDied","Data":"f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60"} Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.578881 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"fe878840-b584-47fb-aca1-bdbdf8f90b43","Type":"ContainerDied","Data":"24fd5a5e2c73d17a2c7b7ce83a3fc6b3dcc6015ed1e2a7fbe4eb45a26d0d8694"} Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.578920 4874 scope.go:117] "RemoveContainer" containerID="f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.579158 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.627522 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.632826 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.645104 4874 scope.go:117] "RemoveContainer" containerID="89564f4dbf9a2d19321b838663656b209f8663a0ed711a85fb6ba3eae0a0748e" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.667297 4874 scope.go:117] "RemoveContainer" containerID="f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60" Jan 22 11:57:44 crc kubenswrapper[4874]: E0122 11:57:44.667706 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60\": container with ID starting with f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60 not found: ID does not exist" containerID="f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.667753 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60"} err="failed to get container status \"f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60\": rpc error: code = NotFound desc = could not find container \"f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60\": container with ID starting with f11f8a3f17d73f5c135bc5c8ab5d78fb1685b97c2cb1d7e7bfa5d902cafa8c60 not found: ID does not exist" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.667780 4874 scope.go:117] "RemoveContainer" containerID="89564f4dbf9a2d19321b838663656b209f8663a0ed711a85fb6ba3eae0a0748e" Jan 22 11:57:44 crc kubenswrapper[4874]: E0122 11:57:44.668163 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89564f4dbf9a2d19321b838663656b209f8663a0ed711a85fb6ba3eae0a0748e\": container with ID starting with 89564f4dbf9a2d19321b838663656b209f8663a0ed711a85fb6ba3eae0a0748e not found: ID does not exist" containerID="89564f4dbf9a2d19321b838663656b209f8663a0ed711a85fb6ba3eae0a0748e" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.668226 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89564f4dbf9a2d19321b838663656b209f8663a0ed711a85fb6ba3eae0a0748e"} err="failed to get container status \"89564f4dbf9a2d19321b838663656b209f8663a0ed711a85fb6ba3eae0a0748e\": rpc error: code = NotFound desc = could not find container \"89564f4dbf9a2d19321b838663656b209f8663a0ed711a85fb6ba3eae0a0748e\": container with ID starting with 89564f4dbf9a2d19321b838663656b209f8663a0ed711a85fb6ba3eae0a0748e not found: ID does not exist" Jan 22 11:57:44 crc kubenswrapper[4874]: I0122 11:57:44.726219 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe878840-b584-47fb-aca1-bdbdf8f90b43" path="/var/lib/kubelet/pods/fe878840-b584-47fb-aca1-bdbdf8f90b43/volumes" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.711745 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-2-build"] Jan 22 11:57:45 crc kubenswrapper[4874]: E0122 11:57:45.712247 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe878840-b584-47fb-aca1-bdbdf8f90b43" containerName="manage-dockerfile" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.712260 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe878840-b584-47fb-aca1-bdbdf8f90b43" containerName="manage-dockerfile" Jan 22 11:57:45 crc kubenswrapper[4874]: E0122 11:57:45.712283 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe878840-b584-47fb-aca1-bdbdf8f90b43" containerName="docker-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.712289 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe878840-b584-47fb-aca1-bdbdf8f90b43" containerName="docker-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.712387 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe878840-b584-47fb-aca1-bdbdf8f90b43" containerName="docker-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.713297 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.715349 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-jmkbp" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.715622 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-sys-config" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.715964 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-ca" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.716418 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-global-ca" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743098 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildcachedir\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743175 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-run\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743218 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743275 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-pull\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743322 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nptvv\" (UniqueName: \"kubernetes.io/projected/d9c696c3-8c17-45a3-93d2-75801fa0bff4-kube-api-access-nptvv\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743387 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743490 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-system-configs\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743520 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743560 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildworkdir\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743713 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-push\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743921 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-root\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.743958 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.745911 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845075 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildworkdir\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845155 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-push\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845227 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-root\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845250 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845276 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildcachedir\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845306 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-run\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845332 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845357 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-pull\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845386 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nptvv\" (UniqueName: \"kubernetes.io/projected/d9c696c3-8c17-45a3-93d2-75801fa0bff4-kube-api-access-nptvv\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845498 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845516 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildworkdir\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845513 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildcachedir\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845559 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845642 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-system-configs\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845687 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.845996 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-root\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.846123 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.846371 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-system-configs\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.846787 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.847074 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-run\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.847303 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.851280 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-pull\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.853897 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-push\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:45 crc kubenswrapper[4874]: I0122 11:57:45.871200 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nptvv\" (UniqueName: \"kubernetes.io/projected/d9c696c3-8c17-45a3-93d2-75801fa0bff4-kube-api-access-nptvv\") pod \"sg-core-2-build\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " pod="service-telemetry/sg-core-2-build" Jan 22 11:57:46 crc kubenswrapper[4874]: I0122 11:57:46.027149 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Jan 22 11:57:46 crc kubenswrapper[4874]: I0122 11:57:46.272261 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Jan 22 11:57:46 crc kubenswrapper[4874]: I0122 11:57:46.594752 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"d9c696c3-8c17-45a3-93d2-75801fa0bff4","Type":"ContainerStarted","Data":"597c44687b703277dcd83cef195a6ae697ec9676e1137316d5c3929b8b80bf65"} Jan 22 11:57:46 crc kubenswrapper[4874]: I0122 11:57:46.594799 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"d9c696c3-8c17-45a3-93d2-75801fa0bff4","Type":"ContainerStarted","Data":"eef6f96df9cbbdb33314fde51a3910ff53d4c339428b3cdc79ffacacbb25cb6e"} Jan 22 11:57:47 crc kubenswrapper[4874]: I0122 11:57:47.603916 4874 generic.go:334] "Generic (PLEG): container finished" podID="d9c696c3-8c17-45a3-93d2-75801fa0bff4" containerID="597c44687b703277dcd83cef195a6ae697ec9676e1137316d5c3929b8b80bf65" exitCode=0 Jan 22 11:57:47 crc kubenswrapper[4874]: I0122 11:57:47.604025 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"d9c696c3-8c17-45a3-93d2-75801fa0bff4","Type":"ContainerDied","Data":"597c44687b703277dcd83cef195a6ae697ec9676e1137316d5c3929b8b80bf65"} Jan 22 11:57:48 crc kubenswrapper[4874]: I0122 11:57:48.611267 4874 generic.go:334] "Generic (PLEG): container finished" podID="d9c696c3-8c17-45a3-93d2-75801fa0bff4" containerID="732e9d64e794a4557ed9a1689a240417d73a781922ccd0e1dfec357d6dc019bd" exitCode=0 Jan 22 11:57:48 crc kubenswrapper[4874]: I0122 11:57:48.611515 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"d9c696c3-8c17-45a3-93d2-75801fa0bff4","Type":"ContainerDied","Data":"732e9d64e794a4557ed9a1689a240417d73a781922ccd0e1dfec357d6dc019bd"} Jan 22 11:57:48 crc kubenswrapper[4874]: I0122 11:57:48.639832 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-2-build_d9c696c3-8c17-45a3-93d2-75801fa0bff4/manage-dockerfile/0.log" Jan 22 11:57:49 crc kubenswrapper[4874]: I0122 11:57:49.620825 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"d9c696c3-8c17-45a3-93d2-75801fa0bff4","Type":"ContainerStarted","Data":"99530f55f4edee2cf39ea2220ebb4c14a93218e16d5a2640d0c5c1ea60a0cb4c"} Jan 22 11:57:49 crc kubenswrapper[4874]: I0122 11:57:49.659291 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-2-build" podStartSLOduration=4.6592577429999995 podStartE2EDuration="4.659257743s" podCreationTimestamp="2026-01-22 11:57:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 11:57:49.650440464 +0000 UTC m=+1043.495511604" watchObservedRunningTime="2026-01-22 11:57:49.659257743 +0000 UTC m=+1043.504328853" Jan 22 11:59:13 crc kubenswrapper[4874]: I0122 11:59:13.520476 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:59:13 crc kubenswrapper[4874]: I0122 11:59:13.521083 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 11:59:43 crc kubenswrapper[4874]: I0122 11:59:43.520607 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 11:59:43 crc kubenswrapper[4874]: I0122 11:59:43.521011 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.148496 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2"] Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.150511 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.155718 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.156007 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.169756 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2"] Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.311546 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b40805b8-a40b-4854-80b3-28916f4f8a43-secret-volume\") pod \"collect-profiles-29484720-8vrb2\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.311651 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4svdz\" (UniqueName: \"kubernetes.io/projected/b40805b8-a40b-4854-80b3-28916f4f8a43-kube-api-access-4svdz\") pod \"collect-profiles-29484720-8vrb2\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.311710 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b40805b8-a40b-4854-80b3-28916f4f8a43-config-volume\") pod \"collect-profiles-29484720-8vrb2\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.413181 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4svdz\" (UniqueName: \"kubernetes.io/projected/b40805b8-a40b-4854-80b3-28916f4f8a43-kube-api-access-4svdz\") pod \"collect-profiles-29484720-8vrb2\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.413316 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b40805b8-a40b-4854-80b3-28916f4f8a43-config-volume\") pod \"collect-profiles-29484720-8vrb2\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.413527 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b40805b8-a40b-4854-80b3-28916f4f8a43-secret-volume\") pod \"collect-profiles-29484720-8vrb2\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.417000 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b40805b8-a40b-4854-80b3-28916f4f8a43-config-volume\") pod \"collect-profiles-29484720-8vrb2\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.431942 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b40805b8-a40b-4854-80b3-28916f4f8a43-secret-volume\") pod \"collect-profiles-29484720-8vrb2\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.443534 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4svdz\" (UniqueName: \"kubernetes.io/projected/b40805b8-a40b-4854-80b3-28916f4f8a43-kube-api-access-4svdz\") pod \"collect-profiles-29484720-8vrb2\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.483000 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:00 crc kubenswrapper[4874]: I0122 12:00:00.724271 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2"] Jan 22 12:00:01 crc kubenswrapper[4874]: I0122 12:00:01.587553 4874 generic.go:334] "Generic (PLEG): container finished" podID="b40805b8-a40b-4854-80b3-28916f4f8a43" containerID="8a25148d438b262cdaed2dfa6ed49fd79b1992854289d94b46eaf3e9b7e53219" exitCode=0 Jan 22 12:00:01 crc kubenswrapper[4874]: I0122 12:00:01.587664 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" event={"ID":"b40805b8-a40b-4854-80b3-28916f4f8a43","Type":"ContainerDied","Data":"8a25148d438b262cdaed2dfa6ed49fd79b1992854289d94b46eaf3e9b7e53219"} Jan 22 12:00:01 crc kubenswrapper[4874]: I0122 12:00:01.587853 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" event={"ID":"b40805b8-a40b-4854-80b3-28916f4f8a43","Type":"ContainerStarted","Data":"68e5718356b4f1c5306fdfdec998346424b75624bf6b9b2e43a9ae045ce98780"} Jan 22 12:00:02 crc kubenswrapper[4874]: I0122 12:00:02.927546 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.046528 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b40805b8-a40b-4854-80b3-28916f4f8a43-secret-volume\") pod \"b40805b8-a40b-4854-80b3-28916f4f8a43\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.046590 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4svdz\" (UniqueName: \"kubernetes.io/projected/b40805b8-a40b-4854-80b3-28916f4f8a43-kube-api-access-4svdz\") pod \"b40805b8-a40b-4854-80b3-28916f4f8a43\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.046616 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b40805b8-a40b-4854-80b3-28916f4f8a43-config-volume\") pod \"b40805b8-a40b-4854-80b3-28916f4f8a43\" (UID: \"b40805b8-a40b-4854-80b3-28916f4f8a43\") " Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.047746 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b40805b8-a40b-4854-80b3-28916f4f8a43-config-volume" (OuterVolumeSpecName: "config-volume") pod "b40805b8-a40b-4854-80b3-28916f4f8a43" (UID: "b40805b8-a40b-4854-80b3-28916f4f8a43"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.053285 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40805b8-a40b-4854-80b3-28916f4f8a43-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b40805b8-a40b-4854-80b3-28916f4f8a43" (UID: "b40805b8-a40b-4854-80b3-28916f4f8a43"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.068647 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b40805b8-a40b-4854-80b3-28916f4f8a43-kube-api-access-4svdz" (OuterVolumeSpecName: "kube-api-access-4svdz") pod "b40805b8-a40b-4854-80b3-28916f4f8a43" (UID: "b40805b8-a40b-4854-80b3-28916f4f8a43"). InnerVolumeSpecName "kube-api-access-4svdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.147879 4874 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b40805b8-a40b-4854-80b3-28916f4f8a43-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.148248 4874 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b40805b8-a40b-4854-80b3-28916f4f8a43-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.148264 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4svdz\" (UniqueName: \"kubernetes.io/projected/b40805b8-a40b-4854-80b3-28916f4f8a43-kube-api-access-4svdz\") on node \"crc\" DevicePath \"\"" Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.601488 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" event={"ID":"b40805b8-a40b-4854-80b3-28916f4f8a43","Type":"ContainerDied","Data":"68e5718356b4f1c5306fdfdec998346424b75624bf6b9b2e43a9ae045ce98780"} Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.601530 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68e5718356b4f1c5306fdfdec998346424b75624bf6b9b2e43a9ae045ce98780" Jan 22 12:00:03 crc kubenswrapper[4874]: I0122 12:00:03.601555 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2" Jan 22 12:00:13 crc kubenswrapper[4874]: I0122 12:00:13.520788 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:00:13 crc kubenswrapper[4874]: I0122 12:00:13.521424 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:00:13 crc kubenswrapper[4874]: I0122 12:00:13.521495 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 12:00:13 crc kubenswrapper[4874]: I0122 12:00:13.522194 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8527e378a489c65991df9b62c1d10fce3a020be8e3c0c8d8d62b128fb0466805"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 12:00:13 crc kubenswrapper[4874]: I0122 12:00:13.522281 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://8527e378a489c65991df9b62c1d10fce3a020be8e3c0c8d8d62b128fb0466805" gracePeriod=600 Jan 22 12:00:14 crc kubenswrapper[4874]: I0122 12:00:14.696168 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="8527e378a489c65991df9b62c1d10fce3a020be8e3c0c8d8d62b128fb0466805" exitCode=0 Jan 22 12:00:14 crc kubenswrapper[4874]: I0122 12:00:14.696355 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"8527e378a489c65991df9b62c1d10fce3a020be8e3c0c8d8d62b128fb0466805"} Jan 22 12:00:14 crc kubenswrapper[4874]: I0122 12:00:14.696595 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"75a7548055039130cae1c6ba4be8efe015a8bd6db752337a976c1be606300781"} Jan 22 12:00:14 crc kubenswrapper[4874]: I0122 12:00:14.696630 4874 scope.go:117] "RemoveContainer" containerID="2554a3567b106d7ded370e89a12bf13dafb3f02d930bcb0aa478a1f4bf2cf32b" Jan 22 12:00:54 crc kubenswrapper[4874]: I0122 12:00:54.128758 4874 patch_prober.go:28] interesting pod/console-operator-58897d9998-mftbv container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 12:00:54 crc kubenswrapper[4874]: I0122 12:00:54.132703 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-mftbv" podUID="8a67b9a8-ad8a-40e3-955c-53aed07a9140" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 22 12:01:24 crc kubenswrapper[4874]: I0122 12:01:24.204781 4874 generic.go:334] "Generic (PLEG): container finished" podID="d9c696c3-8c17-45a3-93d2-75801fa0bff4" containerID="99530f55f4edee2cf39ea2220ebb4c14a93218e16d5a2640d0c5c1ea60a0cb4c" exitCode=0 Jan 22 12:01:24 crc kubenswrapper[4874]: I0122 12:01:24.204896 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"d9c696c3-8c17-45a3-93d2-75801fa0bff4","Type":"ContainerDied","Data":"99530f55f4edee2cf39ea2220ebb4c14a93218e16d5a2640d0c5c1ea60a0cb4c"} Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.512694 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600129 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-node-pullsecrets\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600253 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-blob-cache\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600256 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600300 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-proxy-ca-bundles\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600339 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-push\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600376 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildcachedir\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600450 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-root\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600488 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-system-configs\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600513 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-pull\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600542 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildworkdir\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600576 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-run\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600607 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-ca-bundles\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600644 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nptvv\" (UniqueName: \"kubernetes.io/projected/d9c696c3-8c17-45a3-93d2-75801fa0bff4-kube-api-access-nptvv\") pod \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\" (UID: \"d9c696c3-8c17-45a3-93d2-75801fa0bff4\") " Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.600837 4874 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.603216 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.603259 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.603749 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.603888 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.604552 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.609853 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9c696c3-8c17-45a3-93d2-75801fa0bff4-kube-api-access-nptvv" (OuterVolumeSpecName: "kube-api-access-nptvv") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "kube-api-access-nptvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.610733 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-pull" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-pull") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "builder-dockercfg-jmkbp-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.615201 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-push" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-push") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "builder-dockercfg-jmkbp-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.618360 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.701776 4874 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.701811 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-push\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.701821 4874 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.701829 4874 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.701837 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/d9c696c3-8c17-45a3-93d2-75801fa0bff4-builder-dockercfg-jmkbp-pull\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.701846 4874 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.701854 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.701865 4874 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.701873 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nptvv\" (UniqueName: \"kubernetes.io/projected/d9c696c3-8c17-45a3-93d2-75801fa0bff4-kube-api-access-nptvv\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:25 crc kubenswrapper[4874]: I0122 12:01:25.967585 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:01:26 crc kubenswrapper[4874]: I0122 12:01:26.005238 4874 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:26 crc kubenswrapper[4874]: I0122 12:01:26.224987 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"d9c696c3-8c17-45a3-93d2-75801fa0bff4","Type":"ContainerDied","Data":"eef6f96df9cbbdb33314fde51a3910ff53d4c339428b3cdc79ffacacbb25cb6e"} Jan 22 12:01:26 crc kubenswrapper[4874]: I0122 12:01:26.225045 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eef6f96df9cbbdb33314fde51a3910ff53d4c339428b3cdc79ffacacbb25cb6e" Jan 22 12:01:26 crc kubenswrapper[4874]: I0122 12:01:26.225142 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Jan 22 12:01:28 crc kubenswrapper[4874]: I0122 12:01:28.891028 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "d9c696c3-8c17-45a3-93d2-75801fa0bff4" (UID: "d9c696c3-8c17-45a3-93d2-75801fa0bff4"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:01:28 crc kubenswrapper[4874]: I0122 12:01:28.957127 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d9c696c3-8c17-45a3-93d2-75801fa0bff4-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.307372 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 22 12:01:30 crc kubenswrapper[4874]: E0122 12:01:30.308350 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c696c3-8c17-45a3-93d2-75801fa0bff4" containerName="manage-dockerfile" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.308383 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c696c3-8c17-45a3-93d2-75801fa0bff4" containerName="manage-dockerfile" Jan 22 12:01:30 crc kubenswrapper[4874]: E0122 12:01:30.308451 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40805b8-a40b-4854-80b3-28916f4f8a43" containerName="collect-profiles" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.308469 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40805b8-a40b-4854-80b3-28916f4f8a43" containerName="collect-profiles" Jan 22 12:01:30 crc kubenswrapper[4874]: E0122 12:01:30.308503 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c696c3-8c17-45a3-93d2-75801fa0bff4" containerName="docker-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.308520 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c696c3-8c17-45a3-93d2-75801fa0bff4" containerName="docker-build" Jan 22 12:01:30 crc kubenswrapper[4874]: E0122 12:01:30.308549 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c696c3-8c17-45a3-93d2-75801fa0bff4" containerName="git-clone" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.308565 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c696c3-8c17-45a3-93d2-75801fa0bff4" containerName="git-clone" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.308800 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40805b8-a40b-4854-80b3-28916f4f8a43" containerName="collect-profiles" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.308848 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9c696c3-8c17-45a3-93d2-75801fa0bff4" containerName="docker-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.310244 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.312947 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-global-ca" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.315239 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-ca" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.315537 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-jmkbp" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.315611 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-sys-config" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.340360 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377208 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377290 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377378 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-pull\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377537 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377644 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377692 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377758 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr6mf\" (UniqueName: \"kubernetes.io/projected/800f107b-2838-4231-b18a-b79d55fb8462-kube-api-access-mr6mf\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377799 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377831 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377865 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-push\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377901 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.377931 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479070 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-pull\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479159 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479218 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479252 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479291 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr6mf\" (UniqueName: \"kubernetes.io/projected/800f107b-2838-4231-b18a-b79d55fb8462-kube-api-access-mr6mf\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479326 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479369 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479462 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-push\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479512 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479547 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479613 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479683 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.479972 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.480179 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.480302 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.480582 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.481116 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.481123 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.481928 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.481943 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.481989 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.486697 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-pull\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.488360 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-push\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.514675 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr6mf\" (UniqueName: \"kubernetes.io/projected/800f107b-2838-4231-b18a-b79d55fb8462-kube-api-access-mr6mf\") pod \"sg-bridge-1-build\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.640745 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:30 crc kubenswrapper[4874]: I0122 12:01:30.943761 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 22 12:01:31 crc kubenswrapper[4874]: I0122 12:01:31.271919 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"800f107b-2838-4231-b18a-b79d55fb8462","Type":"ContainerStarted","Data":"a99c233f6fa3271be7db376911a5bd86bbfc74b03e87481eec91c179785d49bb"} Jan 22 12:01:32 crc kubenswrapper[4874]: I0122 12:01:32.290588 4874 generic.go:334] "Generic (PLEG): container finished" podID="800f107b-2838-4231-b18a-b79d55fb8462" containerID="0de26c01cefe3d14fbdb3b1a34fd21d6c52587de7238901147f563699b679d87" exitCode=0 Jan 22 12:01:32 crc kubenswrapper[4874]: I0122 12:01:32.290656 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"800f107b-2838-4231-b18a-b79d55fb8462","Type":"ContainerDied","Data":"0de26c01cefe3d14fbdb3b1a34fd21d6c52587de7238901147f563699b679d87"} Jan 22 12:01:33 crc kubenswrapper[4874]: I0122 12:01:33.300223 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"800f107b-2838-4231-b18a-b79d55fb8462","Type":"ContainerStarted","Data":"bc11c6aeda8d6c6bd381ba9d5974f75e7324fcfb45d919a0312312b2bb9e4228"} Jan 22 12:01:33 crc kubenswrapper[4874]: I0122 12:01:33.336973 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-1-build" podStartSLOduration=3.336954468 podStartE2EDuration="3.336954468s" podCreationTimestamp="2026-01-22 12:01:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 12:01:33.33243415 +0000 UTC m=+1267.177505240" watchObservedRunningTime="2026-01-22 12:01:33.336954468 +0000 UTC m=+1267.182025548" Jan 22 12:01:40 crc kubenswrapper[4874]: I0122 12:01:40.616543 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 22 12:01:40 crc kubenswrapper[4874]: I0122 12:01:40.617119 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/sg-bridge-1-build" podUID="800f107b-2838-4231-b18a-b79d55fb8462" containerName="docker-build" containerID="cri-o://bc11c6aeda8d6c6bd381ba9d5974f75e7324fcfb45d919a0312312b2bb9e4228" gracePeriod=30 Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.366852 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_800f107b-2838-4231-b18a-b79d55fb8462/docker-build/0.log" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.367781 4874 generic.go:334] "Generic (PLEG): container finished" podID="800f107b-2838-4231-b18a-b79d55fb8462" containerID="bc11c6aeda8d6c6bd381ba9d5974f75e7324fcfb45d919a0312312b2bb9e4228" exitCode=1 Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.367813 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"800f107b-2838-4231-b18a-b79d55fb8462","Type":"ContainerDied","Data":"bc11c6aeda8d6c6bd381ba9d5974f75e7324fcfb45d919a0312312b2bb9e4228"} Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.568386 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_800f107b-2838-4231-b18a-b79d55fb8462/docker-build/0.log" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.569057 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.745863 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-push\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.745953 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mr6mf\" (UniqueName: \"kubernetes.io/projected/800f107b-2838-4231-b18a-b79d55fb8462-kube-api-access-mr6mf\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746034 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-ca-bundles\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746077 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-buildcachedir\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746138 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-proxy-ca-bundles\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746180 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-system-configs\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746226 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-pull\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746217 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746319 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-build-blob-cache\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746702 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-node-pullsecrets\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746752 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-buildworkdir\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746789 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-run\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746813 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.746826 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-root\") pod \"800f107b-2838-4231-b18a-b79d55fb8462\" (UID: \"800f107b-2838-4231-b18a-b79d55fb8462\") " Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.747470 4874 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.747495 4874 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/800f107b-2838-4231-b18a-b79d55fb8462-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.747543 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.747589 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.748352 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.748950 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.749959 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.769960 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-pull" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-pull") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "builder-dockercfg-jmkbp-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.770555 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/800f107b-2838-4231-b18a-b79d55fb8462-kube-api-access-mr6mf" (OuterVolumeSpecName: "kube-api-access-mr6mf") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "kube-api-access-mr6mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.771540 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-push" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-push") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "builder-dockercfg-jmkbp-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.845346 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.848309 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-pull\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.848374 4874 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.848390 4874 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.848422 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.848439 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/800f107b-2838-4231-b18a-b79d55fb8462-builder-dockercfg-jmkbp-push\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.848454 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mr6mf\" (UniqueName: \"kubernetes.io/projected/800f107b-2838-4231-b18a-b79d55fb8462-kube-api-access-mr6mf\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.848465 4874 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.848475 4874 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:41 crc kubenswrapper[4874]: I0122 12:01:41.848488 4874 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/800f107b-2838-4231-b18a-b79d55fb8462-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.149530 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "800f107b-2838-4231-b18a-b79d55fb8462" (UID: "800f107b-2838-4231-b18a-b79d55fb8462"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.152829 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/800f107b-2838-4231-b18a-b79d55fb8462-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.264100 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-2-build"] Jan 22 12:01:42 crc kubenswrapper[4874]: E0122 12:01:42.264463 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="800f107b-2838-4231-b18a-b79d55fb8462" containerName="manage-dockerfile" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.264485 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="800f107b-2838-4231-b18a-b79d55fb8462" containerName="manage-dockerfile" Jan 22 12:01:42 crc kubenswrapper[4874]: E0122 12:01:42.264501 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="800f107b-2838-4231-b18a-b79d55fb8462" containerName="docker-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.264510 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="800f107b-2838-4231-b18a-b79d55fb8462" containerName="docker-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.264681 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="800f107b-2838-4231-b18a-b79d55fb8462" containerName="docker-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.265815 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.268281 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-sys-config" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.268527 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-ca" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.268979 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-global-ca" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.278144 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.373421 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_800f107b-2838-4231-b18a-b79d55fb8462/docker-build/0.log" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.374039 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"800f107b-2838-4231-b18a-b79d55fb8462","Type":"ContainerDied","Data":"a99c233f6fa3271be7db376911a5bd86bbfc74b03e87481eec91c179785d49bb"} Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.374097 4874 scope.go:117] "RemoveContainer" containerID="bc11c6aeda8d6c6bd381ba9d5974f75e7324fcfb45d919a0312312b2bb9e4228" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.374132 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.433906 4874 scope.go:117] "RemoveContainer" containerID="0de26c01cefe3d14fbdb3b1a34fd21d6c52587de7238901147f563699b679d87" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.438899 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.447949 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456373 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-push\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456426 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456451 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456468 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456486 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-pull\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456501 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456575 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456591 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456609 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456631 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456651 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.456751 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hfjz\" (UniqueName: \"kubernetes.io/projected/c087bdb2-7a13-44f4-a3e8-d023752790b1-kube-api-access-4hfjz\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.557822 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.557908 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.557947 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.557996 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.558033 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.558022 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.558079 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hfjz\" (UniqueName: \"kubernetes.io/projected/c087bdb2-7a13-44f4-a3e8-d023752790b1-kube-api-access-4hfjz\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.558162 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-push\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.558209 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.558252 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.558283 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.558311 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-pull\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.558346 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.558808 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.558989 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.559500 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.559896 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.560097 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.560242 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.560495 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.561794 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.563599 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-push\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.567074 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-pull\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.590687 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hfjz\" (UniqueName: \"kubernetes.io/projected/c087bdb2-7a13-44f4-a3e8-d023752790b1-kube-api-access-4hfjz\") pod \"sg-bridge-2-build\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.727210 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="800f107b-2838-4231-b18a-b79d55fb8462" path="/var/lib/kubelet/pods/800f107b-2838-4231-b18a-b79d55fb8462/volumes" Jan 22 12:01:42 crc kubenswrapper[4874]: I0122 12:01:42.885440 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Jan 22 12:01:43 crc kubenswrapper[4874]: I0122 12:01:43.153636 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Jan 22 12:01:43 crc kubenswrapper[4874]: I0122 12:01:43.382959 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"c087bdb2-7a13-44f4-a3e8-d023752790b1","Type":"ContainerStarted","Data":"03c145c6ca77e91abe6da3eee50916d0faf5143179dbc4bb6305f384059014cb"} Jan 22 12:01:44 crc kubenswrapper[4874]: I0122 12:01:44.390431 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"c087bdb2-7a13-44f4-a3e8-d023752790b1","Type":"ContainerStarted","Data":"bec974edd7c85b0ec7b6a166376c0bec44230e25fd3b1650be8945a58ee8234b"} Jan 22 12:01:45 crc kubenswrapper[4874]: I0122 12:01:45.400174 4874 generic.go:334] "Generic (PLEG): container finished" podID="c087bdb2-7a13-44f4-a3e8-d023752790b1" containerID="bec974edd7c85b0ec7b6a166376c0bec44230e25fd3b1650be8945a58ee8234b" exitCode=0 Jan 22 12:01:45 crc kubenswrapper[4874]: I0122 12:01:45.400240 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"c087bdb2-7a13-44f4-a3e8-d023752790b1","Type":"ContainerDied","Data":"bec974edd7c85b0ec7b6a166376c0bec44230e25fd3b1650be8945a58ee8234b"} Jan 22 12:01:46 crc kubenswrapper[4874]: I0122 12:01:46.413169 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"c087bdb2-7a13-44f4-a3e8-d023752790b1","Type":"ContainerDied","Data":"23590652507fa8f13559f30aef24b304be84ab519306c12e229704bd157a03c9"} Jan 22 12:01:46 crc kubenswrapper[4874]: I0122 12:01:46.412995 4874 generic.go:334] "Generic (PLEG): container finished" podID="c087bdb2-7a13-44f4-a3e8-d023752790b1" containerID="23590652507fa8f13559f30aef24b304be84ab519306c12e229704bd157a03c9" exitCode=0 Jan 22 12:01:46 crc kubenswrapper[4874]: I0122 12:01:46.475504 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-2-build_c087bdb2-7a13-44f4-a3e8-d023752790b1/manage-dockerfile/0.log" Jan 22 12:01:47 crc kubenswrapper[4874]: I0122 12:01:47.426080 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"c087bdb2-7a13-44f4-a3e8-d023752790b1","Type":"ContainerStarted","Data":"03fce667683d464b8333babfad9386c271cabdec4aa5c2556892d60d9307683d"} Jan 22 12:01:47 crc kubenswrapper[4874]: I0122 12:01:47.475937 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-2-build" podStartSLOduration=5.475909836 podStartE2EDuration="5.475909836s" podCreationTimestamp="2026-01-22 12:01:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 12:01:47.467122266 +0000 UTC m=+1281.312193396" watchObservedRunningTime="2026-01-22 12:01:47.475909836 +0000 UTC m=+1281.320980936" Jan 22 12:02:39 crc kubenswrapper[4874]: I0122 12:02:39.855596 4874 generic.go:334] "Generic (PLEG): container finished" podID="c087bdb2-7a13-44f4-a3e8-d023752790b1" containerID="03fce667683d464b8333babfad9386c271cabdec4aa5c2556892d60d9307683d" exitCode=0 Jan 22 12:02:39 crc kubenswrapper[4874]: I0122 12:02:39.856265 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"c087bdb2-7a13-44f4-a3e8-d023752790b1","Type":"ContainerDied","Data":"03fce667683d464b8333babfad9386c271cabdec4aa5c2556892d60d9307683d"} Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.098330 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177459 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-root\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177539 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-pull\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177585 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-node-pullsecrets\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177623 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-ca-bundles\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177666 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildcachedir\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177700 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildworkdir\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177728 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-blob-cache\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177762 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-push\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177793 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hfjz\" (UniqueName: \"kubernetes.io/projected/c087bdb2-7a13-44f4-a3e8-d023752790b1-kube-api-access-4hfjz\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177829 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-proxy-ca-bundles\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177864 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-run\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.177917 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-system-configs\") pod \"c087bdb2-7a13-44f4-a3e8-d023752790b1\" (UID: \"c087bdb2-7a13-44f4-a3e8-d023752790b1\") " Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.178600 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.178704 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.179557 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.179727 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.180975 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.181640 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.183933 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.184091 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-push" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-push") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "builder-dockercfg-jmkbp-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.184149 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-pull" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-pull") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "builder-dockercfg-jmkbp-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.184206 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c087bdb2-7a13-44f4-a3e8-d023752790b1-kube-api-access-4hfjz" (OuterVolumeSpecName: "kube-api-access-4hfjz") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "kube-api-access-4hfjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.278836 4874 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.278880 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-pull\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.278897 4874 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.278908 4874 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.278919 4874 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.278932 4874 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.278943 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/c087bdb2-7a13-44f4-a3e8-d023752790b1-builder-dockercfg-jmkbp-push\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.278953 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hfjz\" (UniqueName: \"kubernetes.io/projected/c087bdb2-7a13-44f4-a3e8-d023752790b1-kube-api-access-4hfjz\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.278968 4874 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.278979 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.287920 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.380956 4874 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.877446 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"c087bdb2-7a13-44f4-a3e8-d023752790b1","Type":"ContainerDied","Data":"03c145c6ca77e91abe6da3eee50916d0faf5143179dbc4bb6305f384059014cb"} Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.877556 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03c145c6ca77e91abe6da3eee50916d0faf5143179dbc4bb6305f384059014cb" Jan 22 12:02:41 crc kubenswrapper[4874]: I0122 12:02:41.877583 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Jan 22 12:02:42 crc kubenswrapper[4874]: I0122 12:02:42.043008 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "c087bdb2-7a13-44f4-a3e8-d023752790b1" (UID: "c087bdb2-7a13-44f4-a3e8-d023752790b1"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:02:42 crc kubenswrapper[4874]: I0122 12:02:42.091753 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c087bdb2-7a13-44f4-a3e8-d023752790b1-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:43 crc kubenswrapper[4874]: I0122 12:02:43.520040 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:02:43 crc kubenswrapper[4874]: I0122 12:02:43.520124 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.510006 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 22 12:02:45 crc kubenswrapper[4874]: E0122 12:02:45.510549 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c087bdb2-7a13-44f4-a3e8-d023752790b1" containerName="manage-dockerfile" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.510561 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="c087bdb2-7a13-44f4-a3e8-d023752790b1" containerName="manage-dockerfile" Jan 22 12:02:45 crc kubenswrapper[4874]: E0122 12:02:45.510572 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c087bdb2-7a13-44f4-a3e8-d023752790b1" containerName="git-clone" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.510578 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="c087bdb2-7a13-44f4-a3e8-d023752790b1" containerName="git-clone" Jan 22 12:02:45 crc kubenswrapper[4874]: E0122 12:02:45.510593 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c087bdb2-7a13-44f4-a3e8-d023752790b1" containerName="docker-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.510599 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="c087bdb2-7a13-44f4-a3e8-d023752790b1" containerName="docker-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.510692 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="c087bdb2-7a13-44f4-a3e8-d023752790b1" containerName="docker-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.511251 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.519842 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-ca" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.520487 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-sys-config" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.520779 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-jmkbp" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.521608 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-global-ca" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.537599 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642472 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642512 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642530 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642553 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642582 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642605 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642626 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642660 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642678 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642698 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rndv4\" (UniqueName: \"kubernetes.io/projected/d7593be4-689b-4c31-a19d-e92e0532ff7a-kube-api-access-rndv4\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642716 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.642736 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.743746 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.743816 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.743867 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rndv4\" (UniqueName: \"kubernetes.io/projected/d7593be4-689b-4c31-a19d-e92e0532ff7a-kube-api-access-rndv4\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.743903 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.743945 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744049 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744084 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744119 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744158 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744210 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744253 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744298 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744366 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744444 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744481 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744534 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744241 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.744746 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.745076 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.745147 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.745189 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.753136 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.753143 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.763466 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rndv4\" (UniqueName: \"kubernetes.io/projected/d7593be4-689b-4c31-a19d-e92e0532ff7a-kube-api-access-rndv4\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:45 crc kubenswrapper[4874]: I0122 12:02:45.825896 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:46 crc kubenswrapper[4874]: I0122 12:02:46.346390 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 22 12:02:46 crc kubenswrapper[4874]: I0122 12:02:46.936204 4874 generic.go:334] "Generic (PLEG): container finished" podID="d7593be4-689b-4c31-a19d-e92e0532ff7a" containerID="cc01ca7e6056bf37c999ae4b7cc25e8543e1c445265cf7eaf584235b78c8febf" exitCode=0 Jan 22 12:02:46 crc kubenswrapper[4874]: I0122 12:02:46.936287 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"d7593be4-689b-4c31-a19d-e92e0532ff7a","Type":"ContainerDied","Data":"cc01ca7e6056bf37c999ae4b7cc25e8543e1c445265cf7eaf584235b78c8febf"} Jan 22 12:02:46 crc kubenswrapper[4874]: I0122 12:02:46.936378 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"d7593be4-689b-4c31-a19d-e92e0532ff7a","Type":"ContainerStarted","Data":"881208a8d8e40a48481311cf4b043654d5ed9b3946807ac3ddc58f75310baeea"} Jan 22 12:02:47 crc kubenswrapper[4874]: I0122 12:02:47.945290 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"d7593be4-689b-4c31-a19d-e92e0532ff7a","Type":"ContainerStarted","Data":"4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872"} Jan 22 12:02:47 crc kubenswrapper[4874]: I0122 12:02:47.991379 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-1-build" podStartSLOduration=2.991346356 podStartE2EDuration="2.991346356s" podCreationTimestamp="2026-01-22 12:02:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 12:02:47.981209105 +0000 UTC m=+1341.826280225" watchObservedRunningTime="2026-01-22 12:02:47.991346356 +0000 UTC m=+1341.836417466" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.261843 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.262633 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/prometheus-webhook-snmp-1-build" podUID="d7593be4-689b-4c31-a19d-e92e0532ff7a" containerName="docker-build" containerID="cri-o://4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872" gracePeriod=30 Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.675127 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_d7593be4-689b-4c31-a19d-e92e0532ff7a/docker-build/0.log" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.676204 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800296 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildcachedir\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800378 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-root\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800464 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-proxy-ca-bundles\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800510 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-ca-bundles\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800521 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800542 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-push\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800708 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-blob-cache\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800759 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildworkdir\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800809 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rndv4\" (UniqueName: \"kubernetes.io/projected/d7593be4-689b-4c31-a19d-e92e0532ff7a-kube-api-access-rndv4\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800843 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-run\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800897 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-pull\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.800975 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-system-configs\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.801012 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-node-pullsecrets\") pod \"d7593be4-689b-4c31-a19d-e92e0532ff7a\" (UID: \"d7593be4-689b-4c31-a19d-e92e0532ff7a\") " Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.801485 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.801829 4874 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.803860 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.804234 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.804272 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.804260 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.805020 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.810826 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7593be4-689b-4c31-a19d-e92e0532ff7a-kube-api-access-rndv4" (OuterVolumeSpecName: "kube-api-access-rndv4") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "kube-api-access-rndv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.813077 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-pull" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-pull") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "builder-dockercfg-jmkbp-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.818670 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-push" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-push") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "builder-dockercfg-jmkbp-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.878005 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.903367 4874 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.903605 4874 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.903615 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rndv4\" (UniqueName: \"kubernetes.io/projected/d7593be4-689b-4c31-a19d-e92e0532ff7a-kube-api-access-rndv4\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.903625 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.903633 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-pull\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.903642 4874 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.903651 4874 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d7593be4-689b-4c31-a19d-e92e0532ff7a-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.903659 4874 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.903670 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/d7593be4-689b-4c31-a19d-e92e0532ff7a-builder-dockercfg-jmkbp-push\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:56 crc kubenswrapper[4874]: I0122 12:02:56.903677 4874 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7593be4-689b-4c31-a19d-e92e0532ff7a-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.005337 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_d7593be4-689b-4c31-a19d-e92e0532ff7a/docker-build/0.log" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.005884 4874 generic.go:334] "Generic (PLEG): container finished" podID="d7593be4-689b-4c31-a19d-e92e0532ff7a" containerID="4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872" exitCode=1 Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.005924 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"d7593be4-689b-4c31-a19d-e92e0532ff7a","Type":"ContainerDied","Data":"4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872"} Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.005946 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.005954 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"d7593be4-689b-4c31-a19d-e92e0532ff7a","Type":"ContainerDied","Data":"881208a8d8e40a48481311cf4b043654d5ed9b3946807ac3ddc58f75310baeea"} Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.005974 4874 scope.go:117] "RemoveContainer" containerID="4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.037308 4874 scope.go:117] "RemoveContainer" containerID="cc01ca7e6056bf37c999ae4b7cc25e8543e1c445265cf7eaf584235b78c8febf" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.066087 4874 scope.go:117] "RemoveContainer" containerID="4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872" Jan 22 12:02:57 crc kubenswrapper[4874]: E0122 12:02:57.066618 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872\": container with ID starting with 4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872 not found: ID does not exist" containerID="4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.066672 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872"} err="failed to get container status \"4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872\": rpc error: code = NotFound desc = could not find container \"4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872\": container with ID starting with 4838f8e664a6c743d5586c810fc107f1bf65bc6fac610bf0b92e38da2d9d3872 not found: ID does not exist" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.066705 4874 scope.go:117] "RemoveContainer" containerID="cc01ca7e6056bf37c999ae4b7cc25e8543e1c445265cf7eaf584235b78c8febf" Jan 22 12:02:57 crc kubenswrapper[4874]: E0122 12:02:57.067341 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc01ca7e6056bf37c999ae4b7cc25e8543e1c445265cf7eaf584235b78c8febf\": container with ID starting with cc01ca7e6056bf37c999ae4b7cc25e8543e1c445265cf7eaf584235b78c8febf not found: ID does not exist" containerID="cc01ca7e6056bf37c999ae4b7cc25e8543e1c445265cf7eaf584235b78c8febf" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.067382 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc01ca7e6056bf37c999ae4b7cc25e8543e1c445265cf7eaf584235b78c8febf"} err="failed to get container status \"cc01ca7e6056bf37c999ae4b7cc25e8543e1c445265cf7eaf584235b78c8febf\": rpc error: code = NotFound desc = could not find container \"cc01ca7e6056bf37c999ae4b7cc25e8543e1c445265cf7eaf584235b78c8febf\": container with ID starting with cc01ca7e6056bf37c999ae4b7cc25e8543e1c445265cf7eaf584235b78c8febf not found: ID does not exist" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.351376 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "d7593be4-689b-4c31-a19d-e92e0532ff7a" (UID: "d7593be4-689b-4c31-a19d-e92e0532ff7a"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.409223 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d7593be4-689b-4c31-a19d-e92e0532ff7a-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.667480 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.675518 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.904016 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Jan 22 12:02:57 crc kubenswrapper[4874]: E0122 12:02:57.904268 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7593be4-689b-4c31-a19d-e92e0532ff7a" containerName="manage-dockerfile" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.904283 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7593be4-689b-4c31-a19d-e92e0532ff7a" containerName="manage-dockerfile" Jan 22 12:02:57 crc kubenswrapper[4874]: E0122 12:02:57.904298 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7593be4-689b-4c31-a19d-e92e0532ff7a" containerName="docker-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.904307 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7593be4-689b-4c31-a19d-e92e0532ff7a" containerName="docker-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.904455 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7593be4-689b-4c31-a19d-e92e0532ff7a" containerName="docker-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.905413 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.907129 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-global-ca" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.907422 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-jmkbp" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.907879 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-sys-config" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.910719 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-ca" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918602 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918640 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918661 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918693 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918727 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918749 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918779 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918816 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz6zp\" (UniqueName: \"kubernetes.io/projected/0930f68b-3161-43da-80c9-440cf31d98b9-kube-api-access-nz6zp\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918852 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918872 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918907 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.918946 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:57 crc kubenswrapper[4874]: I0122 12:02:57.937178 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.020358 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.020519 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.020565 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.020600 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.020631 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.020679 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.020724 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.020754 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.020832 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.020993 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.021268 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.021424 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz6zp\" (UniqueName: \"kubernetes.io/projected/0930f68b-3161-43da-80c9-440cf31d98b9-kube-api-access-nz6zp\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.021494 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.021585 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.021685 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.021750 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.021789 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.021888 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.022164 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.022163 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.022622 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.024885 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.027094 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.054663 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz6zp\" (UniqueName: \"kubernetes.io/projected/0930f68b-3161-43da-80c9-440cf31d98b9-kube-api-access-nz6zp\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.222570 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.550546 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Jan 22 12:02:58 crc kubenswrapper[4874]: W0122 12:02:58.557545 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0930f68b_3161_43da_80c9_440cf31d98b9.slice/crio-886d9f0c67a3a4c8d14ceb141fd085df974e37449fd7e5219c23a2101e17a594 WatchSource:0}: Error finding container 886d9f0c67a3a4c8d14ceb141fd085df974e37449fd7e5219c23a2101e17a594: Status 404 returned error can't find the container with id 886d9f0c67a3a4c8d14ceb141fd085df974e37449fd7e5219c23a2101e17a594 Jan 22 12:02:58 crc kubenswrapper[4874]: I0122 12:02:58.724623 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7593be4-689b-4c31-a19d-e92e0532ff7a" path="/var/lib/kubelet/pods/d7593be4-689b-4c31-a19d-e92e0532ff7a/volumes" Jan 22 12:02:59 crc kubenswrapper[4874]: I0122 12:02:59.027267 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"0930f68b-3161-43da-80c9-440cf31d98b9","Type":"ContainerStarted","Data":"d5c3e73264960ac0886ae573a444480717c77fc3ce266bc2ffb12054f1170f02"} Jan 22 12:02:59 crc kubenswrapper[4874]: I0122 12:02:59.027333 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"0930f68b-3161-43da-80c9-440cf31d98b9","Type":"ContainerStarted","Data":"886d9f0c67a3a4c8d14ceb141fd085df974e37449fd7e5219c23a2101e17a594"} Jan 22 12:03:00 crc kubenswrapper[4874]: I0122 12:03:00.037788 4874 generic.go:334] "Generic (PLEG): container finished" podID="0930f68b-3161-43da-80c9-440cf31d98b9" containerID="d5c3e73264960ac0886ae573a444480717c77fc3ce266bc2ffb12054f1170f02" exitCode=0 Jan 22 12:03:00 crc kubenswrapper[4874]: I0122 12:03:00.037834 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"0930f68b-3161-43da-80c9-440cf31d98b9","Type":"ContainerDied","Data":"d5c3e73264960ac0886ae573a444480717c77fc3ce266bc2ffb12054f1170f02"} Jan 22 12:03:01 crc kubenswrapper[4874]: I0122 12:03:01.046527 4874 generic.go:334] "Generic (PLEG): container finished" podID="0930f68b-3161-43da-80c9-440cf31d98b9" containerID="ecca9b0c68b8c5dab8b5cbc8e453ab9979273f14bca7c42e50e718db90057883" exitCode=0 Jan 22 12:03:01 crc kubenswrapper[4874]: I0122 12:03:01.047049 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"0930f68b-3161-43da-80c9-440cf31d98b9","Type":"ContainerDied","Data":"ecca9b0c68b8c5dab8b5cbc8e453ab9979273f14bca7c42e50e718db90057883"} Jan 22 12:03:01 crc kubenswrapper[4874]: I0122 12:03:01.099169 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_0930f68b-3161-43da-80c9-440cf31d98b9/manage-dockerfile/0.log" Jan 22 12:03:02 crc kubenswrapper[4874]: I0122 12:03:02.057635 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"0930f68b-3161-43da-80c9-440cf31d98b9","Type":"ContainerStarted","Data":"e02bf90168e7d467aa3f3480ae95da68a8df2550b54ab56b09bd571cb9315ecd"} Jan 22 12:03:02 crc kubenswrapper[4874]: I0122 12:03:02.090156 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-2-build" podStartSLOduration=5.090137498 podStartE2EDuration="5.090137498s" podCreationTimestamp="2026-01-22 12:02:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 12:03:02.08727273 +0000 UTC m=+1355.932343850" watchObservedRunningTime="2026-01-22 12:03:02.090137498 +0000 UTC m=+1355.935208568" Jan 22 12:03:13 crc kubenswrapper[4874]: I0122 12:03:13.520200 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:03:13 crc kubenswrapper[4874]: I0122 12:03:13.520998 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:03:43 crc kubenswrapper[4874]: I0122 12:03:43.520542 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:03:43 crc kubenswrapper[4874]: I0122 12:03:43.521116 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:03:43 crc kubenswrapper[4874]: I0122 12:03:43.521159 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 12:03:43 crc kubenswrapper[4874]: I0122 12:03:43.521691 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"75a7548055039130cae1c6ba4be8efe015a8bd6db752337a976c1be606300781"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 12:03:43 crc kubenswrapper[4874]: I0122 12:03:43.521743 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://75a7548055039130cae1c6ba4be8efe015a8bd6db752337a976c1be606300781" gracePeriod=600 Jan 22 12:03:44 crc kubenswrapper[4874]: I0122 12:03:44.363928 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="75a7548055039130cae1c6ba4be8efe015a8bd6db752337a976c1be606300781" exitCode=0 Jan 22 12:03:44 crc kubenswrapper[4874]: I0122 12:03:44.364048 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"75a7548055039130cae1c6ba4be8efe015a8bd6db752337a976c1be606300781"} Jan 22 12:03:44 crc kubenswrapper[4874]: I0122 12:03:44.364300 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15"} Jan 22 12:03:44 crc kubenswrapper[4874]: I0122 12:03:44.364326 4874 scope.go:117] "RemoveContainer" containerID="8527e378a489c65991df9b62c1d10fce3a020be8e3c0c8d8d62b128fb0466805" Jan 22 12:03:54 crc kubenswrapper[4874]: I0122 12:03:54.436926 4874 generic.go:334] "Generic (PLEG): container finished" podID="0930f68b-3161-43da-80c9-440cf31d98b9" containerID="e02bf90168e7d467aa3f3480ae95da68a8df2550b54ab56b09bd571cb9315ecd" exitCode=0 Jan 22 12:03:54 crc kubenswrapper[4874]: I0122 12:03:54.437068 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"0930f68b-3161-43da-80c9-440cf31d98b9","Type":"ContainerDied","Data":"e02bf90168e7d467aa3f3480ae95da68a8df2550b54ab56b09bd571cb9315ecd"} Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.698831 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.844825 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-system-configs\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.844929 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-ca-bundles\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.844959 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-push\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.845695 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.845767 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.845890 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-buildcachedir\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.845924 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz6zp\" (UniqueName: \"kubernetes.io/projected/0930f68b-3161-43da-80c9-440cf31d98b9-kube-api-access-nz6zp\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.845970 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-pull\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.845981 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.846128 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.846380 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-node-pullsecrets\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.846480 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-buildworkdir\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.846505 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-run\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.846553 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-root\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.846583 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-proxy-ca-bundles\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.846607 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-build-blob-cache\") pod \"0930f68b-3161-43da-80c9-440cf31d98b9\" (UID: \"0930f68b-3161-43da-80c9-440cf31d98b9\") " Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.847182 4874 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.847205 4874 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.847219 4874 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.847231 4874 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0930f68b-3161-43da-80c9-440cf31d98b9-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.847832 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.847943 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.849759 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.851130 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-push" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-push") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "builder-dockercfg-jmkbp-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.854617 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-pull" (OuterVolumeSpecName: "builder-dockercfg-jmkbp-pull") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "builder-dockercfg-jmkbp-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.868692 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0930f68b-3161-43da-80c9-440cf31d98b9-kube-api-access-nz6zp" (OuterVolumeSpecName: "kube-api-access-nz6zp") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "kube-api-access-nz6zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.948639 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-pull\" (UniqueName: \"kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-pull\") on node \"crc\" DevicePath \"\"" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.948676 4874 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.948689 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.948703 4874 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0930f68b-3161-43da-80c9-440cf31d98b9-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.948714 4874 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-jmkbp-push\" (UniqueName: \"kubernetes.io/secret/0930f68b-3161-43da-80c9-440cf31d98b9-builder-dockercfg-jmkbp-push\") on node \"crc\" DevicePath \"\"" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.948727 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz6zp\" (UniqueName: \"kubernetes.io/projected/0930f68b-3161-43da-80c9-440cf31d98b9-kube-api-access-nz6zp\") on node \"crc\" DevicePath \"\"" Jan 22 12:03:55 crc kubenswrapper[4874]: I0122 12:03:55.990047 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:03:56 crc kubenswrapper[4874]: I0122 12:03:56.049845 4874 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 12:03:56 crc kubenswrapper[4874]: I0122 12:03:56.475680 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"0930f68b-3161-43da-80c9-440cf31d98b9","Type":"ContainerDied","Data":"886d9f0c67a3a4c8d14ceb141fd085df974e37449fd7e5219c23a2101e17a594"} Jan 22 12:03:56 crc kubenswrapper[4874]: I0122 12:03:56.475725 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="886d9f0c67a3a4c8d14ceb141fd085df974e37449fd7e5219c23a2101e17a594" Jan 22 12:03:56 crc kubenswrapper[4874]: I0122 12:03:56.475762 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 22 12:03:56 crc kubenswrapper[4874]: I0122 12:03:56.821232 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "0930f68b-3161-43da-80c9-440cf31d98b9" (UID: "0930f68b-3161-43da-80c9-440cf31d98b9"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:03:56 crc kubenswrapper[4874]: I0122 12:03:56.858793 4874 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/0930f68b-3161-43da-80c9-440cf31d98b9-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.661764 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw"] Jan 22 12:04:02 crc kubenswrapper[4874]: E0122 12:04:02.663246 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0930f68b-3161-43da-80c9-440cf31d98b9" containerName="manage-dockerfile" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.663318 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="0930f68b-3161-43da-80c9-440cf31d98b9" containerName="manage-dockerfile" Jan 22 12:04:02 crc kubenswrapper[4874]: E0122 12:04:02.663353 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0930f68b-3161-43da-80c9-440cf31d98b9" containerName="git-clone" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.663485 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="0930f68b-3161-43da-80c9-440cf31d98b9" containerName="git-clone" Jan 22 12:04:02 crc kubenswrapper[4874]: E0122 12:04:02.663574 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0930f68b-3161-43da-80c9-440cf31d98b9" containerName="docker-build" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.663596 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="0930f68b-3161-43da-80c9-440cf31d98b9" containerName="docker-build" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.664064 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="0930f68b-3161-43da-80c9-440cf31d98b9" containerName="docker-build" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.667290 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.670138 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-operator-dockercfg-9gb57" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.670427 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw"] Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.844550 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/f01ecec4-43d6-424e-843b-ef89a732199e-runner\") pod \"smart-gateway-operator-779b7b47dc-4zkhw\" (UID: \"f01ecec4-43d6-424e-843b-ef89a732199e\") " pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.844657 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2tl2\" (UniqueName: \"kubernetes.io/projected/f01ecec4-43d6-424e-843b-ef89a732199e-kube-api-access-c2tl2\") pod \"smart-gateway-operator-779b7b47dc-4zkhw\" (UID: \"f01ecec4-43d6-424e-843b-ef89a732199e\") " pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.945745 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/f01ecec4-43d6-424e-843b-ef89a732199e-runner\") pod \"smart-gateway-operator-779b7b47dc-4zkhw\" (UID: \"f01ecec4-43d6-424e-843b-ef89a732199e\") " pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.945824 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2tl2\" (UniqueName: \"kubernetes.io/projected/f01ecec4-43d6-424e-843b-ef89a732199e-kube-api-access-c2tl2\") pod \"smart-gateway-operator-779b7b47dc-4zkhw\" (UID: \"f01ecec4-43d6-424e-843b-ef89a732199e\") " pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.946311 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/f01ecec4-43d6-424e-843b-ef89a732199e-runner\") pod \"smart-gateway-operator-779b7b47dc-4zkhw\" (UID: \"f01ecec4-43d6-424e-843b-ef89a732199e\") " pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.974013 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2tl2\" (UniqueName: \"kubernetes.io/projected/f01ecec4-43d6-424e-843b-ef89a732199e-kube-api-access-c2tl2\") pod \"smart-gateway-operator-779b7b47dc-4zkhw\" (UID: \"f01ecec4-43d6-424e-843b-ef89a732199e\") " pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" Jan 22 12:04:02 crc kubenswrapper[4874]: I0122 12:04:02.991764 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" Jan 22 12:04:03 crc kubenswrapper[4874]: I0122 12:04:03.194927 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw"] Jan 22 12:04:03 crc kubenswrapper[4874]: I0122 12:04:03.202561 4874 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 12:04:03 crc kubenswrapper[4874]: I0122 12:04:03.520123 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" event={"ID":"f01ecec4-43d6-424e-843b-ef89a732199e","Type":"ContainerStarted","Data":"3cdd36d662ac42074462bc833dfe5a4412991b5028b8df949e8f12fcc10dfb9b"} Jan 22 12:04:07 crc kubenswrapper[4874]: I0122 12:04:07.772666 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-66768bfd46-6v9td"] Jan 22 12:04:07 crc kubenswrapper[4874]: I0122 12:04:07.773876 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-66768bfd46-6v9td" Jan 22 12:04:07 crc kubenswrapper[4874]: I0122 12:04:07.777430 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-operator-dockercfg-hmdt2" Jan 22 12:04:07 crc kubenswrapper[4874]: I0122 12:04:07.790125 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-66768bfd46-6v9td"] Jan 22 12:04:07 crc kubenswrapper[4874]: I0122 12:04:07.819143 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413-runner\") pod \"service-telemetry-operator-66768bfd46-6v9td\" (UID: \"9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413\") " pod="service-telemetry/service-telemetry-operator-66768bfd46-6v9td" Jan 22 12:04:07 crc kubenswrapper[4874]: I0122 12:04:07.819210 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pccd6\" (UniqueName: \"kubernetes.io/projected/9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413-kube-api-access-pccd6\") pod \"service-telemetry-operator-66768bfd46-6v9td\" (UID: \"9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413\") " pod="service-telemetry/service-telemetry-operator-66768bfd46-6v9td" Jan 22 12:04:07 crc kubenswrapper[4874]: I0122 12:04:07.921073 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pccd6\" (UniqueName: \"kubernetes.io/projected/9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413-kube-api-access-pccd6\") pod \"service-telemetry-operator-66768bfd46-6v9td\" (UID: \"9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413\") " pod="service-telemetry/service-telemetry-operator-66768bfd46-6v9td" Jan 22 12:04:07 crc kubenswrapper[4874]: I0122 12:04:07.921179 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413-runner\") pod \"service-telemetry-operator-66768bfd46-6v9td\" (UID: \"9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413\") " pod="service-telemetry/service-telemetry-operator-66768bfd46-6v9td" Jan 22 12:04:07 crc kubenswrapper[4874]: I0122 12:04:07.921653 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413-runner\") pod \"service-telemetry-operator-66768bfd46-6v9td\" (UID: \"9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413\") " pod="service-telemetry/service-telemetry-operator-66768bfd46-6v9td" Jan 22 12:04:07 crc kubenswrapper[4874]: I0122 12:04:07.940026 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pccd6\" (UniqueName: \"kubernetes.io/projected/9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413-kube-api-access-pccd6\") pod \"service-telemetry-operator-66768bfd46-6v9td\" (UID: \"9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413\") " pod="service-telemetry/service-telemetry-operator-66768bfd46-6v9td" Jan 22 12:04:08 crc kubenswrapper[4874]: I0122 12:04:08.095578 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-66768bfd46-6v9td" Jan 22 12:04:13 crc kubenswrapper[4874]: I0122 12:04:13.460750 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-66768bfd46-6v9td"] Jan 22 12:04:15 crc kubenswrapper[4874]: W0122 12:04:15.333207 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c1cf40e_37fa_4e5a_8bb3_8ee8aee9c413.slice/crio-7b271d980e3d4689ef8eb27d1670f7dfb3b1abe89917c0046058f3856f588011 WatchSource:0}: Error finding container 7b271d980e3d4689ef8eb27d1670f7dfb3b1abe89917c0046058f3856f588011: Status 404 returned error can't find the container with id 7b271d980e3d4689ef8eb27d1670f7dfb3b1abe89917c0046058f3856f588011 Jan 22 12:04:15 crc kubenswrapper[4874]: I0122 12:04:15.616088 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-66768bfd46-6v9td" event={"ID":"9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413","Type":"ContainerStarted","Data":"7b271d980e3d4689ef8eb27d1670f7dfb3b1abe89917c0046058f3856f588011"} Jan 22 12:04:18 crc kubenswrapper[4874]: E0122 12:04:18.153013 4874 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/smart-gateway-operator:latest" Jan 22 12:04:18 crc kubenswrapper[4874]: E0122 12:04:18.153278 4874 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/smart-gateway-operator:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:smart-gateway-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:ANSIBLE_VERBOSITY_SMARTGATEWAY_SMARTGATEWAY_INFRA_WATCH,Value:4,ValueFrom:nil,},EnvVar{Name:ANSIBLE_DEBUG_LOGS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CORE_SMARTGATEWAY_IMAGE,Value:image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BRIDGE_SMARTGATEWAY_IMAGE,Value:image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-bridge:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:smart-gateway-operator.v5.0.1769083437,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c2tl2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod smart-gateway-operator-779b7b47dc-4zkhw_service-telemetry(f01ecec4-43d6-424e-843b-ef89a732199e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 22 12:04:18 crc kubenswrapper[4874]: E0122 12:04:18.154634 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" podUID="f01ecec4-43d6-424e-843b-ef89a732199e" Jan 22 12:04:18 crc kubenswrapper[4874]: E0122 12:04:18.649467 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/smart-gateway-operator:latest\\\"\"" pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" podUID="f01ecec4-43d6-424e-843b-ef89a732199e" Jan 22 12:04:23 crc kubenswrapper[4874]: I0122 12:04:23.684069 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-66768bfd46-6v9td" event={"ID":"9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413","Type":"ContainerStarted","Data":"e73c375bccc59c8c66d59408ff25a7a58a9e99a99d320c2eff1c3936b98918b9"} Jan 22 12:04:31 crc kubenswrapper[4874]: I0122 12:04:31.739053 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-66768bfd46-6v9td" podStartSLOduration=17.165570992 podStartE2EDuration="24.73903696s" podCreationTimestamp="2026-01-22 12:04:07 +0000 UTC" firstStartedPulling="2026-01-22 12:04:15.336102844 +0000 UTC m=+1429.181173954" lastFinishedPulling="2026-01-22 12:04:22.909568852 +0000 UTC m=+1436.754639922" observedRunningTime="2026-01-22 12:04:23.712453348 +0000 UTC m=+1437.557524498" watchObservedRunningTime="2026-01-22 12:04:31.73903696 +0000 UTC m=+1445.584108020" Jan 22 12:04:32 crc kubenswrapper[4874]: I0122 12:04:32.752104 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" event={"ID":"f01ecec4-43d6-424e-843b-ef89a732199e","Type":"ContainerStarted","Data":"45613c1b63ae05368aca29c1bd76c66756c2ae4f66666668026c782988c44c8d"} Jan 22 12:04:32 crc kubenswrapper[4874]: I0122 12:04:32.782069 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-779b7b47dc-4zkhw" podStartSLOduration=1.8025797940000001 podStartE2EDuration="30.782041338s" podCreationTimestamp="2026-01-22 12:04:02 +0000 UTC" firstStartedPulling="2026-01-22 12:04:03.202203657 +0000 UTC m=+1417.047274737" lastFinishedPulling="2026-01-22 12:04:32.181665211 +0000 UTC m=+1446.026736281" observedRunningTime="2026-01-22 12:04:32.781301605 +0000 UTC m=+1446.626372675" watchObservedRunningTime="2026-01-22 12:04:32.782041338 +0000 UTC m=+1446.627112418" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.471599 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-lbnpc"] Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.473598 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.478354 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.478490 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.478688 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.479651 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.479693 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.481883 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.483269 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-ghwkd" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.483893 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.484011 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.484037 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-config\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.484064 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-users\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.484083 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.484126 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hv7f\" (UniqueName: \"kubernetes.io/projected/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-kube-api-access-5hv7f\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.484180 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.516081 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-lbnpc"] Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.585071 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.585111 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-config\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.585130 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-users\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.585146 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.585172 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hv7f\" (UniqueName: \"kubernetes.io/projected/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-kube-api-access-5hv7f\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.585205 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.585249 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.586351 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-config\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.595173 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.595209 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.595442 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.596915 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-users\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.600828 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hv7f\" (UniqueName: \"kubernetes.io/projected/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-kube-api-access-5hv7f\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.607850 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-lbnpc\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:45 crc kubenswrapper[4874]: I0122 12:04:45.798299 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:04:46 crc kubenswrapper[4874]: I0122 12:04:46.017689 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-lbnpc"] Jan 22 12:04:46 crc kubenswrapper[4874]: I0122 12:04:46.856470 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" event={"ID":"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527","Type":"ContainerStarted","Data":"991fe91ef8468119983543c90e335d348cd43ff832a854998cadb6f267da781f"} Jan 22 12:05:02 crc kubenswrapper[4874]: I0122 12:05:02.075265 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" event={"ID":"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527","Type":"ContainerStarted","Data":"1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8"} Jan 22 12:05:02 crc kubenswrapper[4874]: I0122 12:05:02.092787 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" podStartSLOduration=2.060106955 podStartE2EDuration="17.092770548s" podCreationTimestamp="2026-01-22 12:04:45 +0000 UTC" firstStartedPulling="2026-01-22 12:04:46.026693631 +0000 UTC m=+1459.871764701" lastFinishedPulling="2026-01-22 12:05:01.059357224 +0000 UTC m=+1474.904428294" observedRunningTime="2026-01-22 12:05:02.090504829 +0000 UTC m=+1475.935575919" watchObservedRunningTime="2026-01-22 12:05:02.092770548 +0000 UTC m=+1475.937841618" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.887457 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-default-0"] Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.888865 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.891634 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-1" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.891771 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-2" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.891764 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"serving-certs-ca-bundle" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.892009 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-stf-dockercfg-bbm2x" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.893333 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-0" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.893372 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-prometheus-proxy-tls" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.893911 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-web-config" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.893919 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.894034 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-tls-assets-0" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.894188 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-session-secret" Jan 22 12:05:03 crc kubenswrapper[4874]: I0122 12:05:03.913873 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052453 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-web-config\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052510 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpdjs\" (UniqueName: \"kubernetes.io/projected/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-kube-api-access-gpdjs\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052530 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052554 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1fec953c-685d-4ed1-bb39-3ba4f9281425\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1fec953c-685d-4ed1-bb39-3ba4f9281425\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052578 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052600 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-config-out\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052626 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-config\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052650 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052695 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-tls-assets\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052711 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052729 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.052754 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154531 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-tls-assets\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154584 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154609 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154643 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154689 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-web-config\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154710 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpdjs\" (UniqueName: \"kubernetes.io/projected/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-kube-api-access-gpdjs\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154730 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154758 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1fec953c-685d-4ed1-bb39-3ba4f9281425\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1fec953c-685d-4ed1-bb39-3ba4f9281425\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154784 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154808 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-config-out\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154837 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-config\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.154868 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: E0122 12:05:04.154993 4874 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Jan 22 12:05:04 crc kubenswrapper[4874]: E0122 12:05:04.155054 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-prometheus-proxy-tls podName:7cd2b3c5-32a6-49c1-984c-86aa8fe36f85 nodeName:}" failed. No retries permitted until 2026-01-22 12:05:04.655033446 +0000 UTC m=+1478.500104516 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "7cd2b3c5-32a6-49c1-984c-86aa8fe36f85") : secret "default-prometheus-proxy-tls" not found Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.156360 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.160960 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.162248 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.165176 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-config-out\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.165177 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.165794 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.167140 4874 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.167176 4874 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1fec953c-685d-4ed1-bb39-3ba4f9281425\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1fec953c-685d-4ed1-bb39-3ba4f9281425\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/37635c52ffc8d45540ce6f928b3d95e2edf1cbb912893c9ece4ebfa760a97525/globalmount\"" pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.170031 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-tls-assets\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.170652 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-config\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.180889 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-web-config\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.186556 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpdjs\" (UniqueName: \"kubernetes.io/projected/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-kube-api-access-gpdjs\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.197286 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1fec953c-685d-4ed1-bb39-3ba4f9281425\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1fec953c-685d-4ed1-bb39-3ba4f9281425\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: I0122 12:05:04.661194 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:04 crc kubenswrapper[4874]: E0122 12:05:04.661567 4874 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Jan 22 12:05:04 crc kubenswrapper[4874]: E0122 12:05:04.661682 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-prometheus-proxy-tls podName:7cd2b3c5-32a6-49c1-984c-86aa8fe36f85 nodeName:}" failed. No retries permitted until 2026-01-22 12:05:05.661658709 +0000 UTC m=+1479.506729779 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "7cd2b3c5-32a6-49c1-984c-86aa8fe36f85") : secret "default-prometheus-proxy-tls" not found Jan 22 12:05:05 crc kubenswrapper[4874]: I0122 12:05:05.675618 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:05 crc kubenswrapper[4874]: I0122 12:05:05.680312 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/7cd2b3c5-32a6-49c1-984c-86aa8fe36f85-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85\") " pod="service-telemetry/prometheus-default-0" Jan 22 12:05:05 crc kubenswrapper[4874]: I0122 12:05:05.755907 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Jan 22 12:05:06 crc kubenswrapper[4874]: I0122 12:05:06.205739 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Jan 22 12:05:07 crc kubenswrapper[4874]: I0122 12:05:07.112610 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85","Type":"ContainerStarted","Data":"5076195ba1893a2766c22408320072c928f35c9d00d20fa815b8b8aac6394ed7"} Jan 22 12:05:11 crc kubenswrapper[4874]: I0122 12:05:11.142845 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85","Type":"ContainerStarted","Data":"46eae4037523c1a86bc383e7393949e1f079f41fa015304765a32a52db657fe6"} Jan 22 12:05:13 crc kubenswrapper[4874]: I0122 12:05:13.483577 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-9qcrx"] Jan 22 12:05:13 crc kubenswrapper[4874]: I0122 12:05:13.485236 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6856cfb745-9qcrx" Jan 22 12:05:13 crc kubenswrapper[4874]: I0122 12:05:13.493444 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-9qcrx"] Jan 22 12:05:13 crc kubenswrapper[4874]: I0122 12:05:13.600094 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzbgl\" (UniqueName: \"kubernetes.io/projected/a14f7a5d-1819-4ca2-8683-96a338c70df6-kube-api-access-fzbgl\") pod \"default-snmp-webhook-6856cfb745-9qcrx\" (UID: \"a14f7a5d-1819-4ca2-8683-96a338c70df6\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-9qcrx" Jan 22 12:05:13 crc kubenswrapper[4874]: I0122 12:05:13.702220 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzbgl\" (UniqueName: \"kubernetes.io/projected/a14f7a5d-1819-4ca2-8683-96a338c70df6-kube-api-access-fzbgl\") pod \"default-snmp-webhook-6856cfb745-9qcrx\" (UID: \"a14f7a5d-1819-4ca2-8683-96a338c70df6\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-9qcrx" Jan 22 12:05:13 crc kubenswrapper[4874]: I0122 12:05:13.724572 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzbgl\" (UniqueName: \"kubernetes.io/projected/a14f7a5d-1819-4ca2-8683-96a338c70df6-kube-api-access-fzbgl\") pod \"default-snmp-webhook-6856cfb745-9qcrx\" (UID: \"a14f7a5d-1819-4ca2-8683-96a338c70df6\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-9qcrx" Jan 22 12:05:13 crc kubenswrapper[4874]: I0122 12:05:13.803486 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6856cfb745-9qcrx" Jan 22 12:05:14 crc kubenswrapper[4874]: I0122 12:05:14.234497 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-9qcrx"] Jan 22 12:05:15 crc kubenswrapper[4874]: I0122 12:05:15.174135 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6856cfb745-9qcrx" event={"ID":"a14f7a5d-1819-4ca2-8683-96a338c70df6","Type":"ContainerStarted","Data":"f407a60fc8f6ab5d4e4b4e0d71fe4be0fc91e3f5e46fbf4568e32dbadd495ea8"} Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.123072 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/alertmanager-default-0"] Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.125256 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.128841 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.128897 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-cluster-tls-config" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.128966 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-alertmanager-proxy-tls" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.129147 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-tls-assets-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.129223 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-stf-dockercfg-cxk5p" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.129166 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-web-config" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.129509 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-generated" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.254159 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.254218 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-web-config\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.254242 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-config-volume\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.254261 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.254287 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-adf40950-4a83-4124-a213-cde377d20e8e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-adf40950-4a83-4124-a213-cde377d20e8e\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.254350 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/30722d93-5804-4aba-a6e3-e2b891356163-tls-assets\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.254370 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.254422 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5fjb\" (UniqueName: \"kubernetes.io/projected/30722d93-5804-4aba-a6e3-e2b891356163-kube-api-access-z5fjb\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.254450 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/30722d93-5804-4aba-a6e3-e2b891356163-config-out\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.355637 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/30722d93-5804-4aba-a6e3-e2b891356163-tls-assets\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.355700 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.355758 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5fjb\" (UniqueName: \"kubernetes.io/projected/30722d93-5804-4aba-a6e3-e2b891356163-kube-api-access-z5fjb\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.355804 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/30722d93-5804-4aba-a6e3-e2b891356163-config-out\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.355850 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.355885 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-web-config\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.355915 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-config-volume\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.355941 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.355977 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-adf40950-4a83-4124-a213-cde377d20e8e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-adf40950-4a83-4124-a213-cde377d20e8e\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: E0122 12:05:17.357992 4874 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Jan 22 12:05:17 crc kubenswrapper[4874]: E0122 12:05:17.358248 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls podName:30722d93-5804-4aba-a6e3-e2b891356163 nodeName:}" failed. No retries permitted until 2026-01-22 12:05:17.858229398 +0000 UTC m=+1491.703300478 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "30722d93-5804-4aba-a6e3-e2b891356163") : secret "default-alertmanager-proxy-tls" not found Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.362970 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/30722d93-5804-4aba-a6e3-e2b891356163-config-out\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.363274 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-config-volume\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.363368 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-web-config\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.363735 4874 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.363767 4874 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-adf40950-4a83-4124-a213-cde377d20e8e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-adf40950-4a83-4124-a213-cde377d20e8e\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2d7c76d617ad7234da60527edf901d0a9ef34c3c89379835b82ca1c2da10a6ce/globalmount\"" pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.364313 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.371167 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.375038 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5fjb\" (UniqueName: \"kubernetes.io/projected/30722d93-5804-4aba-a6e3-e2b891356163-kube-api-access-z5fjb\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.376828 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/30722d93-5804-4aba-a6e3-e2b891356163-tls-assets\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.412458 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-adf40950-4a83-4124-a213-cde377d20e8e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-adf40950-4a83-4124-a213-cde377d20e8e\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: E0122 12:05:17.522797 4874 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7cd2b3c5_32a6_49c1_984c_86aa8fe36f85.slice/crio-conmon-46eae4037523c1a86bc383e7393949e1f079f41fa015304765a32a52db657fe6.scope\": RecentStats: unable to find data in memory cache]" Jan 22 12:05:17 crc kubenswrapper[4874]: I0122 12:05:17.865738 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:17 crc kubenswrapper[4874]: E0122 12:05:17.865913 4874 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Jan 22 12:05:17 crc kubenswrapper[4874]: E0122 12:05:17.865980 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls podName:30722d93-5804-4aba-a6e3-e2b891356163 nodeName:}" failed. No retries permitted until 2026-01-22 12:05:18.865965625 +0000 UTC m=+1492.711036695 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "30722d93-5804-4aba-a6e3-e2b891356163") : secret "default-alertmanager-proxy-tls" not found Jan 22 12:05:18 crc kubenswrapper[4874]: I0122 12:05:18.199901 4874 generic.go:334] "Generic (PLEG): container finished" podID="7cd2b3c5-32a6-49c1-984c-86aa8fe36f85" containerID="46eae4037523c1a86bc383e7393949e1f079f41fa015304765a32a52db657fe6" exitCode=0 Jan 22 12:05:18 crc kubenswrapper[4874]: I0122 12:05:18.199992 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85","Type":"ContainerDied","Data":"46eae4037523c1a86bc383e7393949e1f079f41fa015304765a32a52db657fe6"} Jan 22 12:05:18 crc kubenswrapper[4874]: I0122 12:05:18.881182 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:18 crc kubenswrapper[4874]: E0122 12:05:18.881485 4874 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Jan 22 12:05:18 crc kubenswrapper[4874]: E0122 12:05:18.881623 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls podName:30722d93-5804-4aba-a6e3-e2b891356163 nodeName:}" failed. No retries permitted until 2026-01-22 12:05:20.881591343 +0000 UTC m=+1494.726662453 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "30722d93-5804-4aba-a6e3-e2b891356163") : secret "default-alertmanager-proxy-tls" not found Jan 22 12:05:20 crc kubenswrapper[4874]: I0122 12:05:20.957689 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:20 crc kubenswrapper[4874]: I0122 12:05:20.977583 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/30722d93-5804-4aba-a6e3-e2b891356163-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"30722d93-5804-4aba-a6e3-e2b891356163\") " pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:21 crc kubenswrapper[4874]: I0122 12:05:21.043257 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Jan 22 12:05:23 crc kubenswrapper[4874]: I0122 12:05:23.384089 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Jan 22 12:05:24 crc kubenswrapper[4874]: I0122 12:05:24.245291 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"30722d93-5804-4aba-a6e3-e2b891356163","Type":"ContainerStarted","Data":"ec150f9482f77e1f4afd4ad1558d86ac0e170cb8e3de262b1ba2df809cf9e66d"} Jan 22 12:05:24 crc kubenswrapper[4874]: I0122 12:05:24.247289 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6856cfb745-9qcrx" event={"ID":"a14f7a5d-1819-4ca2-8683-96a338c70df6","Type":"ContainerStarted","Data":"791b86629f1493b42f6ca4eeb5ea835f85573ce790df81d91166f2f18796dbdb"} Jan 22 12:05:24 crc kubenswrapper[4874]: I0122 12:05:24.260629 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-snmp-webhook-6856cfb745-9qcrx" podStartSLOduration=2.271680537 podStartE2EDuration="11.260605771s" podCreationTimestamp="2026-01-22 12:05:13 +0000 UTC" firstStartedPulling="2026-01-22 12:05:14.246798024 +0000 UTC m=+1488.091869104" lastFinishedPulling="2026-01-22 12:05:23.235723268 +0000 UTC m=+1497.080794338" observedRunningTime="2026-01-22 12:05:24.259481476 +0000 UTC m=+1498.104552556" watchObservedRunningTime="2026-01-22 12:05:24.260605771 +0000 UTC m=+1498.105676851" Jan 22 12:05:26 crc kubenswrapper[4874]: I0122 12:05:26.266470 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"30722d93-5804-4aba-a6e3-e2b891356163","Type":"ContainerStarted","Data":"7ac51e4a91f65653b5b9e0030d22b5ffb622d7862ef9dad4a669c63b8bfcd8b2"} Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.190387 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg"] Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.193539 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.195633 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-coll-meter-proxy-tls" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.197771 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-dockercfg-jzxhb" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.198239 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-session-secret" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.202765 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg"] Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.203106 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-meter-sg-core-configmap" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.312475 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85","Type":"ContainerStarted","Data":"19ec2555e9c2cfd5b72886151b286bbd34c73b723b47e1ce79a93b2d5ff57b3d"} Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.313904 4874 generic.go:334] "Generic (PLEG): container finished" podID="30722d93-5804-4aba-a6e3-e2b891356163" containerID="7ac51e4a91f65653b5b9e0030d22b5ffb622d7862ef9dad4a669c63b8bfcd8b2" exitCode=0 Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.313954 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"30722d93-5804-4aba-a6e3-e2b891356163","Type":"ContainerDied","Data":"7ac51e4a91f65653b5b9e0030d22b5ffb622d7862ef9dad4a669c63b8bfcd8b2"} Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.346325 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zbzw\" (UniqueName: \"kubernetes.io/projected/16eb2f20-a170-4504-8601-a08d7174edef-kube-api-access-9zbzw\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.346449 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/16eb2f20-a170-4504-8601-a08d7174edef-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.346503 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.346546 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/16eb2f20-a170-4504-8601-a08d7174edef-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.346574 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.448434 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zbzw\" (UniqueName: \"kubernetes.io/projected/16eb2f20-a170-4504-8601-a08d7174edef-kube-api-access-9zbzw\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.448600 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/16eb2f20-a170-4504-8601-a08d7174edef-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.448657 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.448697 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/16eb2f20-a170-4504-8601-a08d7174edef-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.448721 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: E0122 12:05:33.450559 4874 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Jan 22 12:05:33 crc kubenswrapper[4874]: E0122 12:05:33.450642 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-default-cloud1-coll-meter-proxy-tls podName:16eb2f20-a170-4504-8601-a08d7174edef nodeName:}" failed. No retries permitted until 2026-01-22 12:05:33.950617371 +0000 UTC m=+1507.795688561 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" (UID: "16eb2f20-a170-4504-8601-a08d7174edef") : secret "default-cloud1-coll-meter-proxy-tls" not found Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.451271 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/16eb2f20-a170-4504-8601-a08d7174edef-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.451959 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/16eb2f20-a170-4504-8601-a08d7174edef-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.457143 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.467034 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zbzw\" (UniqueName: \"kubernetes.io/projected/16eb2f20-a170-4504-8601-a08d7174edef-kube-api-access-9zbzw\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: I0122 12:05:33.959116 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:33 crc kubenswrapper[4874]: E0122 12:05:33.959322 4874 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Jan 22 12:05:33 crc kubenswrapper[4874]: E0122 12:05:33.959644 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-default-cloud1-coll-meter-proxy-tls podName:16eb2f20-a170-4504-8601-a08d7174edef nodeName:}" failed. No retries permitted until 2026-01-22 12:05:34.959626147 +0000 UTC m=+1508.804697217 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" (UID: "16eb2f20-a170-4504-8601-a08d7174edef") : secret "default-cloud1-coll-meter-proxy-tls" not found Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.759553 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np"] Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.760976 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.762979 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-ceil-meter-proxy-tls" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.763379 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-meter-sg-core-configmap" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.773662 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np"] Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.871900 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/1cca42d2-5390-4dc0-90fd-6abef972f828-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.871983 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/1cca42d2-5390-4dc0-90fd-6abef972f828-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.872009 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blszq\" (UniqueName: \"kubernetes.io/projected/1cca42d2-5390-4dc0-90fd-6abef972f828-kube-api-access-blszq\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.872036 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.872074 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.973666 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blszq\" (UniqueName: \"kubernetes.io/projected/1cca42d2-5390-4dc0-90fd-6abef972f828-kube-api-access-blszq\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.973719 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.973756 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.973811 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/1cca42d2-5390-4dc0-90fd-6abef972f828-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.973844 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.973867 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/1cca42d2-5390-4dc0-90fd-6abef972f828-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: E0122 12:05:34.973919 4874 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 22 12:05:34 crc kubenswrapper[4874]: E0122 12:05:34.974033 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-default-cloud1-ceil-meter-proxy-tls podName:1cca42d2-5390-4dc0-90fd-6abef972f828 nodeName:}" failed. No retries permitted until 2026-01-22 12:05:35.474010128 +0000 UTC m=+1509.319081258 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" (UID: "1cca42d2-5390-4dc0-90fd-6abef972f828") : secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.974444 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/1cca42d2-5390-4dc0-90fd-6abef972f828-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.976055 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/1cca42d2-5390-4dc0-90fd-6abef972f828-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.980014 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.990311 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/16eb2f20-a170-4504-8601-a08d7174edef-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg\" (UID: \"16eb2f20-a170-4504-8601-a08d7174edef\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:34 crc kubenswrapper[4874]: I0122 12:05:34.990626 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blszq\" (UniqueName: \"kubernetes.io/projected/1cca42d2-5390-4dc0-90fd-6abef972f828-kube-api-access-blszq\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:35 crc kubenswrapper[4874]: I0122 12:05:35.010247 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" Jan 22 12:05:35 crc kubenswrapper[4874]: I0122 12:05:35.330147 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85","Type":"ContainerStarted","Data":"c332ddf8b4c4c5266dfde74289d4bb221b6529600fba655a0d5bd085cb49f109"} Jan 22 12:05:35 crc kubenswrapper[4874]: I0122 12:05:35.480337 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:35 crc kubenswrapper[4874]: E0122 12:05:35.480592 4874 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 22 12:05:35 crc kubenswrapper[4874]: E0122 12:05:35.480663 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-default-cloud1-ceil-meter-proxy-tls podName:1cca42d2-5390-4dc0-90fd-6abef972f828 nodeName:}" failed. No retries permitted until 2026-01-22 12:05:36.480648511 +0000 UTC m=+1510.325719581 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" (UID: "1cca42d2-5390-4dc0-90fd-6abef972f828") : secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 22 12:05:35 crc kubenswrapper[4874]: I0122 12:05:35.745456 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg"] Jan 22 12:05:35 crc kubenswrapper[4874]: W0122 12:05:35.754991 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16eb2f20_a170_4504_8601_a08d7174edef.slice/crio-6faf6bdfd8162b5f04b23a8a351b6dc2e9b82adf41e742fc72f0f9ec7dfb2565 WatchSource:0}: Error finding container 6faf6bdfd8162b5f04b23a8a351b6dc2e9b82adf41e742fc72f0f9ec7dfb2565: Status 404 returned error can't find the container with id 6faf6bdfd8162b5f04b23a8a351b6dc2e9b82adf41e742fc72f0f9ec7dfb2565 Jan 22 12:05:36 crc kubenswrapper[4874]: I0122 12:05:36.340609 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"30722d93-5804-4aba-a6e3-e2b891356163","Type":"ContainerStarted","Data":"b34999e847ede69071117228fd59d6a319bb4137fa841753278e8c7325cd0548"} Jan 22 12:05:36 crc kubenswrapper[4874]: I0122 12:05:36.342544 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" event={"ID":"16eb2f20-a170-4504-8601-a08d7174edef","Type":"ContainerStarted","Data":"6faf6bdfd8162b5f04b23a8a351b6dc2e9b82adf41e742fc72f0f9ec7dfb2565"} Jan 22 12:05:36 crc kubenswrapper[4874]: I0122 12:05:36.499302 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:36 crc kubenswrapper[4874]: I0122 12:05:36.526229 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1cca42d2-5390-4dc0-90fd-6abef972f828-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np\" (UID: \"1cca42d2-5390-4dc0-90fd-6abef972f828\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:36 crc kubenswrapper[4874]: I0122 12:05:36.638859 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" Jan 22 12:05:37 crc kubenswrapper[4874]: I0122 12:05:37.091346 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np"] Jan 22 12:05:38 crc kubenswrapper[4874]: I0122 12:05:38.359043 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" event={"ID":"1cca42d2-5390-4dc0-90fd-6abef972f828","Type":"ContainerStarted","Data":"7568bf2ee43f93d9bedbe3888875a2a9b8f779219dcb356fa148ce2b8472a51f"} Jan 22 12:05:38 crc kubenswrapper[4874]: I0122 12:05:38.362713 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"30722d93-5804-4aba-a6e3-e2b891356163","Type":"ContainerStarted","Data":"aa5ea25f32a76dff4abbac5521051de55d0791e87c40f78904cd5d5e3cc27020"} Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.373262 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf"] Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.375103 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.377033 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-sens-meter-proxy-tls" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.377239 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-sens-meter-sg-core-configmap" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.386366 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf"] Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.487329 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.487392 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gqt4\" (UniqueName: \"kubernetes.io/projected/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-kube-api-access-9gqt4\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.487427 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.487575 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.487723 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.590187 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.590287 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.590326 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gqt4\" (UniqueName: \"kubernetes.io/projected/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-kube-api-access-9gqt4\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.590349 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.590372 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.590775 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: E0122 12:05:41.590861 4874 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Jan 22 12:05:41 crc kubenswrapper[4874]: E0122 12:05:41.590910 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-default-cloud1-sens-meter-proxy-tls podName:7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2 nodeName:}" failed. No retries permitted until 2026-01-22 12:05:42.090894827 +0000 UTC m=+1515.935965897 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" (UID: "7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2") : secret "default-cloud1-sens-meter-proxy-tls" not found Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.595619 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.600312 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:41 crc kubenswrapper[4874]: I0122 12:05:41.609318 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gqt4\" (UniqueName: \"kubernetes.io/projected/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-kube-api-access-9gqt4\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:42 crc kubenswrapper[4874]: I0122 12:05:42.104710 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:42 crc kubenswrapper[4874]: E0122 12:05:42.104871 4874 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Jan 22 12:05:42 crc kubenswrapper[4874]: E0122 12:05:42.104943 4874 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-default-cloud1-sens-meter-proxy-tls podName:7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2 nodeName:}" failed. No retries permitted until 2026-01-22 12:05:43.104923147 +0000 UTC m=+1516.949994217 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" (UID: "7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2") : secret "default-cloud1-sens-meter-proxy-tls" not found Jan 22 12:05:43 crc kubenswrapper[4874]: I0122 12:05:43.116977 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:43 crc kubenswrapper[4874]: I0122 12:05:43.131023 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf\" (UID: \"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:43 crc kubenswrapper[4874]: I0122 12:05:43.220829 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" Jan 22 12:05:43 crc kubenswrapper[4874]: I0122 12:05:43.422867 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"7cd2b3c5-32a6-49c1-984c-86aa8fe36f85","Type":"ContainerStarted","Data":"f4f8e2b71fca6a5ae2ec4b19e5944e57282e21c8ee6397e3f04a26746c9e3e6c"} Jan 22 12:05:43 crc kubenswrapper[4874]: I0122 12:05:43.424847 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" event={"ID":"16eb2f20-a170-4504-8601-a08d7174edef","Type":"ContainerStarted","Data":"bf6822cb867a1bc23e35dd1ff5500debdbf8238e8a8d1db6bcf22724dfff1ed1"} Jan 22 12:05:43 crc kubenswrapper[4874]: I0122 12:05:43.426586 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" event={"ID":"1cca42d2-5390-4dc0-90fd-6abef972f828","Type":"ContainerStarted","Data":"b6d7d181d3e7422af148da1ee32a5cbfb5c7d8afd452164cfdc65eba9cf1f192"} Jan 22 12:05:43 crc kubenswrapper[4874]: I0122 12:05:43.431448 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"30722d93-5804-4aba-a6e3-e2b891356163","Type":"ContainerStarted","Data":"47c2747d7fd53f2fd0cd892766f3e53d7a763817adfc97900b0db150b45a8c99"} Jan 22 12:05:43 crc kubenswrapper[4874]: I0122 12:05:43.449391 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-default-0" podStartSLOduration=5.034048635 podStartE2EDuration="41.449379757s" podCreationTimestamp="2026-01-22 12:05:02 +0000 UTC" firstStartedPulling="2026-01-22 12:05:06.212427755 +0000 UTC m=+1480.057498825" lastFinishedPulling="2026-01-22 12:05:42.627758877 +0000 UTC m=+1516.472829947" observedRunningTime="2026-01-22 12:05:43.444294711 +0000 UTC m=+1517.289365781" watchObservedRunningTime="2026-01-22 12:05:43.449379757 +0000 UTC m=+1517.294450827" Jan 22 12:05:43 crc kubenswrapper[4874]: I0122 12:05:43.476694 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/alertmanager-default-0" podStartSLOduration=18.166342296 podStartE2EDuration="27.476674074s" podCreationTimestamp="2026-01-22 12:05:16 +0000 UTC" firstStartedPulling="2026-01-22 12:05:33.31651269 +0000 UTC m=+1507.161583780" lastFinishedPulling="2026-01-22 12:05:42.626844488 +0000 UTC m=+1516.471915558" observedRunningTime="2026-01-22 12:05:43.472444774 +0000 UTC m=+1517.317515834" watchObservedRunningTime="2026-01-22 12:05:43.476674074 +0000 UTC m=+1517.321745144" Jan 22 12:05:43 crc kubenswrapper[4874]: I0122 12:05:43.520813 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:05:43 crc kubenswrapper[4874]: I0122 12:05:43.521071 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:05:44 crc kubenswrapper[4874]: I0122 12:05:44.052385 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf"] Jan 22 12:05:44 crc kubenswrapper[4874]: W0122 12:05:44.077915 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c0b7de5_b0a4_4c4d_93c3_1feba19a95a2.slice/crio-df982c53fcb5f37bf927450c8ea79cec679ef0f5cf6513841079b6ef4e4578e2 WatchSource:0}: Error finding container df982c53fcb5f37bf927450c8ea79cec679ef0f5cf6513841079b6ef4e4578e2: Status 404 returned error can't find the container with id df982c53fcb5f37bf927450c8ea79cec679ef0f5cf6513841079b6ef4e4578e2 Jan 22 12:05:44 crc kubenswrapper[4874]: I0122 12:05:44.441294 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" event={"ID":"1cca42d2-5390-4dc0-90fd-6abef972f828","Type":"ContainerStarted","Data":"0b8c4a0b0ba7c449f8d752c598d619f6497f68d22e47bcf468a644699c35b9f8"} Jan 22 12:05:44 crc kubenswrapper[4874]: I0122 12:05:44.442081 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" event={"ID":"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2","Type":"ContainerStarted","Data":"df982c53fcb5f37bf927450c8ea79cec679ef0f5cf6513841079b6ef4e4578e2"} Jan 22 12:05:44 crc kubenswrapper[4874]: I0122 12:05:44.446724 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" event={"ID":"16eb2f20-a170-4504-8601-a08d7174edef","Type":"ContainerStarted","Data":"d00586b99f40c23c83e98d796f708a92dd64728a1c6fc574bfa7124a09972b2f"} Jan 22 12:05:45 crc kubenswrapper[4874]: I0122 12:05:45.462044 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" event={"ID":"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2","Type":"ContainerStarted","Data":"461c8f650e5a8a977666075e1277e3c9267495a8c7840f9acdf0c5d6a6800f08"} Jan 22 12:05:45 crc kubenswrapper[4874]: I0122 12:05:45.462313 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" event={"ID":"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2","Type":"ContainerStarted","Data":"cc0d66438daa34f1ff09fdd821cd6a828285cdf4e2c301e2734ce23de7c377f5"} Jan 22 12:05:45 crc kubenswrapper[4874]: I0122 12:05:45.756593 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/prometheus-default-0" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.037363 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7"] Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.039428 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.044194 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-cert" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.044473 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-event-sg-core-configmap" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.059243 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7"] Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.116546 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8df9n\" (UniqueName: \"kubernetes.io/projected/79746e6a-bee0-42eb-840c-36075d04abde-kube-api-access-8df9n\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.116600 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/79746e6a-bee0-42eb-840c-36075d04abde-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.116627 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/79746e6a-bee0-42eb-840c-36075d04abde-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.116695 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/79746e6a-bee0-42eb-840c-36075d04abde-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.217435 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/79746e6a-bee0-42eb-840c-36075d04abde-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.217502 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8df9n\" (UniqueName: \"kubernetes.io/projected/79746e6a-bee0-42eb-840c-36075d04abde-kube-api-access-8df9n\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.217528 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/79746e6a-bee0-42eb-840c-36075d04abde-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.217546 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/79746e6a-bee0-42eb-840c-36075d04abde-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.218391 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/79746e6a-bee0-42eb-840c-36075d04abde-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.218682 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/79746e6a-bee0-42eb-840c-36075d04abde-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.227214 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/79746e6a-bee0-42eb-840c-36075d04abde-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.232440 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8df9n\" (UniqueName: \"kubernetes.io/projected/79746e6a-bee0-42eb-840c-36075d04abde-kube-api-access-8df9n\") pod \"default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7\" (UID: \"79746e6a-bee0-42eb-840c-36075d04abde\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:48 crc kubenswrapper[4874]: I0122 12:05:48.365557 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.069161 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9"] Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.070739 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.074515 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-event-sg-core-configmap" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.081202 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9"] Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.234169 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwsdz\" (UniqueName: \"kubernetes.io/projected/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-kube-api-access-fwsdz\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.234230 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.234277 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.234319 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.245232 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7"] Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.335874 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.335984 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwsdz\" (UniqueName: \"kubernetes.io/projected/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-kube-api-access-fwsdz\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.336061 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.336166 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.336454 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.337810 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.341624 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.358353 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwsdz\" (UniqueName: \"kubernetes.io/projected/01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3-kube-api-access-fwsdz\") pod \"default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9\" (UID: \"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.395765 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.492763 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" event={"ID":"16eb2f20-a170-4504-8601-a08d7174edef","Type":"ContainerStarted","Data":"69c433e99a725517200fb9d45ed6e81da3887cc6894fd645f13ecc0bb6daffba"} Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.497177 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" event={"ID":"1cca42d2-5390-4dc0-90fd-6abef972f828","Type":"ContainerStarted","Data":"f83992e9db1ba9c89ed7185fc4772855270eace8775a5e26f14083ad85ef4497"} Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.499733 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" event={"ID":"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2","Type":"ContainerStarted","Data":"0fc95eff3e5747b70e69ba8740b6d2f7899fec0e29b7a5271a2cd615230c1863"} Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.503156 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" event={"ID":"79746e6a-bee0-42eb-840c-36075d04abde","Type":"ContainerStarted","Data":"57ae4c13120d233f1b05b1010c95ceb5a571f302e4e0502b5e01d31027df9bca"} Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.525927 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" podStartSLOduration=3.384248665 podStartE2EDuration="16.52590654s" podCreationTimestamp="2026-01-22 12:05:33 +0000 UTC" firstStartedPulling="2026-01-22 12:05:35.758025475 +0000 UTC m=+1509.603096535" lastFinishedPulling="2026-01-22 12:05:48.89968333 +0000 UTC m=+1522.744754410" observedRunningTime="2026-01-22 12:05:49.515441779 +0000 UTC m=+1523.360512849" watchObservedRunningTime="2026-01-22 12:05:49.52590654 +0000 UTC m=+1523.370977610" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.546002 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" podStartSLOduration=4.10197025 podStartE2EDuration="15.545984045s" podCreationTimestamp="2026-01-22 12:05:34 +0000 UTC" firstStartedPulling="2026-01-22 12:05:37.507490743 +0000 UTC m=+1511.352561813" lastFinishedPulling="2026-01-22 12:05:48.951504538 +0000 UTC m=+1522.796575608" observedRunningTime="2026-01-22 12:05:49.544705606 +0000 UTC m=+1523.389776676" watchObservedRunningTime="2026-01-22 12:05:49.545984045 +0000 UTC m=+1523.391055115" Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.572290 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" podStartSLOduration=3.736024075 podStartE2EDuration="8.572269722s" podCreationTimestamp="2026-01-22 12:05:41 +0000 UTC" firstStartedPulling="2026-01-22 12:05:44.082505408 +0000 UTC m=+1517.927576478" lastFinishedPulling="2026-01-22 12:05:48.918751055 +0000 UTC m=+1522.763822125" observedRunningTime="2026-01-22 12:05:49.564254875 +0000 UTC m=+1523.409325945" watchObservedRunningTime="2026-01-22 12:05:49.572269722 +0000 UTC m=+1523.417340792" Jan 22 12:05:49 crc kubenswrapper[4874]: W0122 12:05:49.847964 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01e2e1fb_0e3b_4c3b_b5c5_df459233c6e3.slice/crio-d5d4378713e394dc5a7268b56220f34dbea522db178c42678d9c82528beccdb1 WatchSource:0}: Error finding container d5d4378713e394dc5a7268b56220f34dbea522db178c42678d9c82528beccdb1: Status 404 returned error can't find the container with id d5d4378713e394dc5a7268b56220f34dbea522db178c42678d9c82528beccdb1 Jan 22 12:05:49 crc kubenswrapper[4874]: I0122 12:05:49.850801 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9"] Jan 22 12:05:50 crc kubenswrapper[4874]: I0122 12:05:50.511592 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" event={"ID":"79746e6a-bee0-42eb-840c-36075d04abde","Type":"ContainerStarted","Data":"09cb582dc76bac620de55013695a6049eb521a343c611714eaf77b7038c60d7e"} Jan 22 12:05:50 crc kubenswrapper[4874]: I0122 12:05:50.512856 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" event={"ID":"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3","Type":"ContainerStarted","Data":"d5d4378713e394dc5a7268b56220f34dbea522db178c42678d9c82528beccdb1"} Jan 22 12:05:51 crc kubenswrapper[4874]: I0122 12:05:51.021877 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/prometheus-default-0" Jan 22 12:05:51 crc kubenswrapper[4874]: I0122 12:05:51.055902 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/prometheus-default-0" Jan 22 12:05:51 crc kubenswrapper[4874]: I0122 12:05:51.520020 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" event={"ID":"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3","Type":"ContainerStarted","Data":"0451bbcc35433072c924780c0fd1a6e8ba140eace530b10301001354bb9c5c0f"} Jan 22 12:05:51 crc kubenswrapper[4874]: I0122 12:05:51.521796 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" event={"ID":"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3","Type":"ContainerStarted","Data":"514e40ce0ac94f7f5ee03f53f502f032cb0df43f43b4da1487e85d3afe4ac50a"} Jan 22 12:05:51 crc kubenswrapper[4874]: I0122 12:05:51.521855 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" event={"ID":"79746e6a-bee0-42eb-840c-36075d04abde","Type":"ContainerStarted","Data":"58c31f117d1242b5994f1205597c4e1e95231093a0ea8abb96f2a73941f5bccc"} Jan 22 12:05:51 crc kubenswrapper[4874]: I0122 12:05:51.541608 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" podStartSLOduration=1.102316962 podStartE2EDuration="2.541105054s" podCreationTimestamp="2026-01-22 12:05:49 +0000 UTC" firstStartedPulling="2026-01-22 12:05:49.852981217 +0000 UTC m=+1523.698052297" lastFinishedPulling="2026-01-22 12:05:51.291769319 +0000 UTC m=+1525.136840389" observedRunningTime="2026-01-22 12:05:51.537947557 +0000 UTC m=+1525.383018657" watchObservedRunningTime="2026-01-22 12:05:51.541105054 +0000 UTC m=+1525.386176124" Jan 22 12:05:51 crc kubenswrapper[4874]: I0122 12:05:51.565104 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" podStartSLOduration=2.446574867 podStartE2EDuration="3.56508416s" podCreationTimestamp="2026-01-22 12:05:48 +0000 UTC" firstStartedPulling="2026-01-22 12:05:49.238023943 +0000 UTC m=+1523.083095013" lastFinishedPulling="2026-01-22 12:05:50.356533236 +0000 UTC m=+1524.201604306" observedRunningTime="2026-01-22 12:05:51.5543129 +0000 UTC m=+1525.399383990" watchObservedRunningTime="2026-01-22 12:05:51.56508416 +0000 UTC m=+1525.410155230" Jan 22 12:05:51 crc kubenswrapper[4874]: I0122 12:05:51.593785 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/prometheus-default-0" Jan 22 12:06:04 crc kubenswrapper[4874]: I0122 12:06:04.891186 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/qdr-test"] Jan 22 12:06:04 crc kubenswrapper[4874]: I0122 12:06:04.893841 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Jan 22 12:06:04 crc kubenswrapper[4874]: I0122 12:06:04.896772 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-selfsigned" Jan 22 12:06:04 crc kubenswrapper[4874]: I0122 12:06:04.896953 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"qdr-test-config" Jan 22 12:06:04 crc kubenswrapper[4874]: I0122 12:06:04.905941 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.016664 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/056ff179-9e03-4f48-b3d3-fa995d0c4093-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"056ff179-9e03-4f48-b3d3-fa995d0c4093\") " pod="service-telemetry/qdr-test" Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.016757 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/056ff179-9e03-4f48-b3d3-fa995d0c4093-qdr-test-config\") pod \"qdr-test\" (UID: \"056ff179-9e03-4f48-b3d3-fa995d0c4093\") " pod="service-telemetry/qdr-test" Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.016834 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjrpk\" (UniqueName: \"kubernetes.io/projected/056ff179-9e03-4f48-b3d3-fa995d0c4093-kube-api-access-rjrpk\") pod \"qdr-test\" (UID: \"056ff179-9e03-4f48-b3d3-fa995d0c4093\") " pod="service-telemetry/qdr-test" Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.117902 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/056ff179-9e03-4f48-b3d3-fa995d0c4093-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"056ff179-9e03-4f48-b3d3-fa995d0c4093\") " pod="service-telemetry/qdr-test" Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.117958 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/056ff179-9e03-4f48-b3d3-fa995d0c4093-qdr-test-config\") pod \"qdr-test\" (UID: \"056ff179-9e03-4f48-b3d3-fa995d0c4093\") " pod="service-telemetry/qdr-test" Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.118017 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjrpk\" (UniqueName: \"kubernetes.io/projected/056ff179-9e03-4f48-b3d3-fa995d0c4093-kube-api-access-rjrpk\") pod \"qdr-test\" (UID: \"056ff179-9e03-4f48-b3d3-fa995d0c4093\") " pod="service-telemetry/qdr-test" Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.118863 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/056ff179-9e03-4f48-b3d3-fa995d0c4093-qdr-test-config\") pod \"qdr-test\" (UID: \"056ff179-9e03-4f48-b3d3-fa995d0c4093\") " pod="service-telemetry/qdr-test" Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.130611 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/056ff179-9e03-4f48-b3d3-fa995d0c4093-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"056ff179-9e03-4f48-b3d3-fa995d0c4093\") " pod="service-telemetry/qdr-test" Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.140136 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjrpk\" (UniqueName: \"kubernetes.io/projected/056ff179-9e03-4f48-b3d3-fa995d0c4093-kube-api-access-rjrpk\") pod \"qdr-test\" (UID: \"056ff179-9e03-4f48-b3d3-fa995d0c4093\") " pod="service-telemetry/qdr-test" Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.215100 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.702879 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.912315 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-lbnpc"] Jan 22 12:06:05 crc kubenswrapper[4874]: I0122 12:06:05.912643 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" podUID="c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" containerName="default-interconnect" containerID="cri-o://1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8" gracePeriod=30 Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.260799 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.438764 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hv7f\" (UniqueName: \"kubernetes.io/projected/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-kube-api-access-5hv7f\") pod \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.438812 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-users\") pod \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.438902 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-config\") pod \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.438965 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-credentials\") pod \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.438992 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-ca\") pod \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.439542 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-config" (OuterVolumeSpecName: "sasl-config") pod "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" (UID: "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527"). InnerVolumeSpecName "sasl-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.439689 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-ca\") pod \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.439772 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-credentials\") pod \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\" (UID: \"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527\") " Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.440079 4874 reconciler_common.go:293] "Volume detached for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-config\") on node \"crc\" DevicePath \"\"" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.444979 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-kube-api-access-5hv7f" (OuterVolumeSpecName: "kube-api-access-5hv7f") pod "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" (UID: "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527"). InnerVolumeSpecName "kube-api-access-5hv7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.445517 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-ca" (OuterVolumeSpecName: "default-interconnect-openstack-ca") pod "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" (UID: "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527"). InnerVolumeSpecName "default-interconnect-openstack-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.445541 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-credentials" (OuterVolumeSpecName: "default-interconnect-openstack-credentials") pod "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" (UID: "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527"). InnerVolumeSpecName "default-interconnect-openstack-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.445610 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-users" (OuterVolumeSpecName: "sasl-users") pod "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" (UID: "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527"). InnerVolumeSpecName "sasl-users". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.445749 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-ca" (OuterVolumeSpecName: "default-interconnect-inter-router-ca") pod "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" (UID: "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527"). InnerVolumeSpecName "default-interconnect-inter-router-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.461649 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-credentials" (OuterVolumeSpecName: "default-interconnect-inter-router-credentials") pod "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" (UID: "c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527"). InnerVolumeSpecName "default-interconnect-inter-router-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.541626 4874 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-credentials\") on node \"crc\" DevicePath \"\"" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.541696 4874 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-inter-router-ca\") on node \"crc\" DevicePath \"\"" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.541726 4874 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-ca\") on node \"crc\" DevicePath \"\"" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.541753 4874 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-default-interconnect-openstack-credentials\") on node \"crc\" DevicePath \"\"" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.541780 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hv7f\" (UniqueName: \"kubernetes.io/projected/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-kube-api-access-5hv7f\") on node \"crc\" DevicePath \"\"" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.541805 4874 reconciler_common.go:293] "Volume detached for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527-sasl-users\") on node \"crc\" DevicePath \"\"" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.624526 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"056ff179-9e03-4f48-b3d3-fa995d0c4093","Type":"ContainerStarted","Data":"bd28960a826427f4fcb49d523ac7fbd479766cbb2296e3ca24891bfcf1c6705e"} Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.627158 4874 generic.go:334] "Generic (PLEG): container finished" podID="01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3" containerID="514e40ce0ac94f7f5ee03f53f502f032cb0df43f43b4da1487e85d3afe4ac50a" exitCode=0 Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.627216 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" event={"ID":"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3","Type":"ContainerDied","Data":"514e40ce0ac94f7f5ee03f53f502f032cb0df43f43b4da1487e85d3afe4ac50a"} Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.627673 4874 scope.go:117] "RemoveContainer" containerID="514e40ce0ac94f7f5ee03f53f502f032cb0df43f43b4da1487e85d3afe4ac50a" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.642341 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2" containerID="461c8f650e5a8a977666075e1277e3c9267495a8c7840f9acdf0c5d6a6800f08" exitCode=0 Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.642438 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" event={"ID":"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2","Type":"ContainerDied","Data":"461c8f650e5a8a977666075e1277e3c9267495a8c7840f9acdf0c5d6a6800f08"} Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.643024 4874 scope.go:117] "RemoveContainer" containerID="461c8f650e5a8a977666075e1277e3c9267495a8c7840f9acdf0c5d6a6800f08" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.662669 4874 generic.go:334] "Generic (PLEG): container finished" podID="79746e6a-bee0-42eb-840c-36075d04abde" containerID="09cb582dc76bac620de55013695a6049eb521a343c611714eaf77b7038c60d7e" exitCode=0 Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.662766 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" event={"ID":"79746e6a-bee0-42eb-840c-36075d04abde","Type":"ContainerDied","Data":"09cb582dc76bac620de55013695a6049eb521a343c611714eaf77b7038c60d7e"} Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.663358 4874 scope.go:117] "RemoveContainer" containerID="09cb582dc76bac620de55013695a6049eb521a343c611714eaf77b7038c60d7e" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.667213 4874 generic.go:334] "Generic (PLEG): container finished" podID="c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" containerID="1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8" exitCode=0 Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.667257 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.667257 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" event={"ID":"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527","Type":"ContainerDied","Data":"1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8"} Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.667375 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-lbnpc" event={"ID":"c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527","Type":"ContainerDied","Data":"991fe91ef8468119983543c90e335d348cd43ff832a854998cadb6f267da781f"} Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.667419 4874 scope.go:117] "RemoveContainer" containerID="1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.671990 4874 generic.go:334] "Generic (PLEG): container finished" podID="16eb2f20-a170-4504-8601-a08d7174edef" containerID="d00586b99f40c23c83e98d796f708a92dd64728a1c6fc574bfa7124a09972b2f" exitCode=0 Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.672041 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" event={"ID":"16eb2f20-a170-4504-8601-a08d7174edef","Type":"ContainerDied","Data":"d00586b99f40c23c83e98d796f708a92dd64728a1c6fc574bfa7124a09972b2f"} Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.672714 4874 scope.go:117] "RemoveContainer" containerID="d00586b99f40c23c83e98d796f708a92dd64728a1c6fc574bfa7124a09972b2f" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.707429 4874 scope.go:117] "RemoveContainer" containerID="1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8" Jan 22 12:06:06 crc kubenswrapper[4874]: E0122 12:06:06.707878 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8\": container with ID starting with 1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8 not found: ID does not exist" containerID="1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.707923 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8"} err="failed to get container status \"1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8\": rpc error: code = NotFound desc = could not find container \"1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8\": container with ID starting with 1942e90d87d0c8ebb015abda0e958e1d2138e35879200aafa6962b056430edd8 not found: ID does not exist" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.768967 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-lbnpc"] Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.777035 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-lbnpc"] Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.864184 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vq2h8"] Jan 22 12:06:06 crc kubenswrapper[4874]: E0122 12:06:06.864510 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" containerName="default-interconnect" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.864525 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" containerName="default-interconnect" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.864692 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" containerName="default-interconnect" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.865550 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:06 crc kubenswrapper[4874]: I0122 12:06:06.872154 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vq2h8"] Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.007065 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-jcvt5"] Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.007957 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.012545 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-ghwkd" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.012709 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.012879 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.016740 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.017353 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.017507 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.017649 4874 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.025129 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-jcvt5"] Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.053536 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj82v\" (UniqueName: \"kubernetes.io/projected/be315a0a-e88a-4a39-b705-b826a4285fc2-kube-api-access-cj82v\") pod \"community-operators-vq2h8\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.053603 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-catalog-content\") pod \"community-operators-vq2h8\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.053734 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-utilities\") pod \"community-operators-vq2h8\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155035 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/da6dd65e-8b70-41a1-8a38-83392c65aef3-sasl-config\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155084 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-catalog-content\") pod \"community-operators-vq2h8\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155108 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-sasl-users\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155136 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155166 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tjnb\" (UniqueName: \"kubernetes.io/projected/da6dd65e-8b70-41a1-8a38-83392c65aef3-kube-api-access-6tjnb\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155197 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155218 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-utilities\") pod \"community-operators-vq2h8\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155253 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155271 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj82v\" (UniqueName: \"kubernetes.io/projected/be315a0a-e88a-4a39-b705-b826a4285fc2-kube-api-access-cj82v\") pod \"community-operators-vq2h8\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155301 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155715 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-catalog-content\") pod \"community-operators-vq2h8\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.155939 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-utilities\") pod \"community-operators-vq2h8\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.174683 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj82v\" (UniqueName: \"kubernetes.io/projected/be315a0a-e88a-4a39-b705-b826a4285fc2-kube-api-access-cj82v\") pod \"community-operators-vq2h8\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.223945 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.256106 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.256151 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/da6dd65e-8b70-41a1-8a38-83392c65aef3-sasl-config\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.256173 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-sasl-users\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.256196 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.256230 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tjnb\" (UniqueName: \"kubernetes.io/projected/da6dd65e-8b70-41a1-8a38-83392c65aef3-kube-api-access-6tjnb\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.256268 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.256309 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.258219 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/da6dd65e-8b70-41a1-8a38-83392c65aef3-sasl-config\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.259193 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-sasl-users\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.262936 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.276006 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.276046 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.293090 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/da6dd65e-8b70-41a1-8a38-83392c65aef3-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.293127 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tjnb\" (UniqueName: \"kubernetes.io/projected/da6dd65e-8b70-41a1-8a38-83392c65aef3-kube-api-access-6tjnb\") pod \"default-interconnect-68864d46cb-jcvt5\" (UID: \"da6dd65e-8b70-41a1-8a38-83392c65aef3\") " pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.342048 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.686232 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" event={"ID":"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3","Type":"ContainerStarted","Data":"699a19bb9b06b8eb15d6f8ce1695ec22e8bc5543fd8a9c4aaee428742ce39ca8"} Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.698208 4874 generic.go:334] "Generic (PLEG): container finished" podID="1cca42d2-5390-4dc0-90fd-6abef972f828" containerID="0b8c4a0b0ba7c449f8d752c598d619f6497f68d22e47bcf468a644699c35b9f8" exitCode=0 Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.698368 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" event={"ID":"1cca42d2-5390-4dc0-90fd-6abef972f828","Type":"ContainerDied","Data":"0b8c4a0b0ba7c449f8d752c598d619f6497f68d22e47bcf468a644699c35b9f8"} Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.699346 4874 scope.go:117] "RemoveContainer" containerID="0b8c4a0b0ba7c449f8d752c598d619f6497f68d22e47bcf468a644699c35b9f8" Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.703802 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" event={"ID":"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2","Type":"ContainerStarted","Data":"8871bcfaa155a6fc80df16ab23ea84232653fb97737ee28e90257e9bd644fd42"} Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.709109 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" event={"ID":"79746e6a-bee0-42eb-840c-36075d04abde","Type":"ContainerStarted","Data":"241fe00fc340b80eefba0f92d3bdf67c1dc9e3306e497d8b33550b4e18947a08"} Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.724886 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" event={"ID":"16eb2f20-a170-4504-8601-a08d7174edef","Type":"ContainerStarted","Data":"9950cfa7893dfcae9ef6566159771b0e333f0fd846d8eadc4a654a109358eb87"} Jan 22 12:06:07 crc kubenswrapper[4874]: W0122 12:06:07.801473 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe315a0a_e88a_4a39_b705_b826a4285fc2.slice/crio-6095ec06d7054068d7a98b6e4ccf292ac930e508cf3c276df0c4600afa1fd83e WatchSource:0}: Error finding container 6095ec06d7054068d7a98b6e4ccf292ac930e508cf3c276df0c4600afa1fd83e: Status 404 returned error can't find the container with id 6095ec06d7054068d7a98b6e4ccf292ac930e508cf3c276df0c4600afa1fd83e Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.805597 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vq2h8"] Jan 22 12:06:07 crc kubenswrapper[4874]: I0122 12:06:07.905994 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-jcvt5"] Jan 22 12:06:08 crc kubenswrapper[4874]: E0122 12:06:08.268379 4874 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01e2e1fb_0e3b_4c3b_b5c5_df459233c6e3.slice/crio-conmon-699a19bb9b06b8eb15d6f8ce1695ec22e8bc5543fd8a9c4aaee428742ce39ca8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c0b7de5_b0a4_4c4d_93c3_1feba19a95a2.slice/crio-8871bcfaa155a6fc80df16ab23ea84232653fb97737ee28e90257e9bd644fd42.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01e2e1fb_0e3b_4c3b_b5c5_df459233c6e3.slice/crio-699a19bb9b06b8eb15d6f8ce1695ec22e8bc5543fd8a9c4aaee428742ce39ca8.scope\": RecentStats: unable to find data in memory cache]" Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.726568 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527" path="/var/lib/kubelet/pods/c4d0bf4d-4f56-41b0-ab4a-a6603bbe0527/volumes" Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.747051 4874 generic.go:334] "Generic (PLEG): container finished" podID="01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3" containerID="699a19bb9b06b8eb15d6f8ce1695ec22e8bc5543fd8a9c4aaee428742ce39ca8" exitCode=0 Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.747121 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" event={"ID":"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3","Type":"ContainerDied","Data":"699a19bb9b06b8eb15d6f8ce1695ec22e8bc5543fd8a9c4aaee428742ce39ca8"} Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.747153 4874 scope.go:117] "RemoveContainer" containerID="514e40ce0ac94f7f5ee03f53f502f032cb0df43f43b4da1487e85d3afe4ac50a" Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.747713 4874 scope.go:117] "RemoveContainer" containerID="699a19bb9b06b8eb15d6f8ce1695ec22e8bc5543fd8a9c4aaee428742ce39ca8" Jan 22 12:06:08 crc kubenswrapper[4874]: E0122 12:06:08.747983 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9_service-telemetry(01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" podUID="01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3" Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.752928 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" event={"ID":"da6dd65e-8b70-41a1-8a38-83392c65aef3","Type":"ContainerStarted","Data":"e1c4dc04409f8689fa42459366ed59cdfc73e546395757896efc9a0f219b7438"} Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.752965 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" event={"ID":"da6dd65e-8b70-41a1-8a38-83392c65aef3","Type":"ContainerStarted","Data":"b080c40c0d5d91f79b4bd8a142e5f2cdb0629430058eed1d125dc0190ff3db3e"} Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.756429 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2" containerID="8871bcfaa155a6fc80df16ab23ea84232653fb97737ee28e90257e9bd644fd42" exitCode=0 Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.756479 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" event={"ID":"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2","Type":"ContainerDied","Data":"8871bcfaa155a6fc80df16ab23ea84232653fb97737ee28e90257e9bd644fd42"} Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.756838 4874 scope.go:117] "RemoveContainer" containerID="8871bcfaa155a6fc80df16ab23ea84232653fb97737ee28e90257e9bd644fd42" Jan 22 12:06:08 crc kubenswrapper[4874]: E0122 12:06:08.756999 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf_service-telemetry(7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" podUID="7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2" Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.758924 4874 generic.go:334] "Generic (PLEG): container finished" podID="79746e6a-bee0-42eb-840c-36075d04abde" containerID="241fe00fc340b80eefba0f92d3bdf67c1dc9e3306e497d8b33550b4e18947a08" exitCode=0 Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.758965 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" event={"ID":"79746e6a-bee0-42eb-840c-36075d04abde","Type":"ContainerDied","Data":"241fe00fc340b80eefba0f92d3bdf67c1dc9e3306e497d8b33550b4e18947a08"} Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.759259 4874 scope.go:117] "RemoveContainer" containerID="241fe00fc340b80eefba0f92d3bdf67c1dc9e3306e497d8b33550b4e18947a08" Jan 22 12:06:08 crc kubenswrapper[4874]: E0122 12:06:08.759449 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7_service-telemetry(79746e6a-bee0-42eb-840c-36075d04abde)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" podUID="79746e6a-bee0-42eb-840c-36075d04abde" Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.787236 4874 generic.go:334] "Generic (PLEG): container finished" podID="be315a0a-e88a-4a39-b705-b826a4285fc2" containerID="4119410dea713ff391706990d3ca3df3ffd65b7a0e739319efe9d22e1aefc65b" exitCode=0 Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.787439 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq2h8" event={"ID":"be315a0a-e88a-4a39-b705-b826a4285fc2","Type":"ContainerDied","Data":"4119410dea713ff391706990d3ca3df3ffd65b7a0e739319efe9d22e1aefc65b"} Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.787637 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq2h8" event={"ID":"be315a0a-e88a-4a39-b705-b826a4285fc2","Type":"ContainerStarted","Data":"6095ec06d7054068d7a98b6e4ccf292ac930e508cf3c276df0c4600afa1fd83e"} Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.800674 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-jcvt5" podStartSLOduration=3.800658912 podStartE2EDuration="3.800658912s" podCreationTimestamp="2026-01-22 12:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 12:06:08.797837826 +0000 UTC m=+1542.642908896" watchObservedRunningTime="2026-01-22 12:06:08.800658912 +0000 UTC m=+1542.645729982" Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.849977 4874 scope.go:117] "RemoveContainer" containerID="461c8f650e5a8a977666075e1277e3c9267495a8c7840f9acdf0c5d6a6800f08" Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.850153 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" event={"ID":"1cca42d2-5390-4dc0-90fd-6abef972f828","Type":"ContainerStarted","Data":"03a5f9d616cc942c97a12adda95ac1fb8d538715f65c06b4d599618c1bca3dc2"} Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.865782 4874 generic.go:334] "Generic (PLEG): container finished" podID="16eb2f20-a170-4504-8601-a08d7174edef" containerID="9950cfa7893dfcae9ef6566159771b0e333f0fd846d8eadc4a654a109358eb87" exitCode=0 Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.865826 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" event={"ID":"16eb2f20-a170-4504-8601-a08d7174edef","Type":"ContainerDied","Data":"9950cfa7893dfcae9ef6566159771b0e333f0fd846d8eadc4a654a109358eb87"} Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.866298 4874 scope.go:117] "RemoveContainer" containerID="9950cfa7893dfcae9ef6566159771b0e333f0fd846d8eadc4a654a109358eb87" Jan 22 12:06:08 crc kubenswrapper[4874]: E0122 12:06:08.866523 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg_service-telemetry(16eb2f20-a170-4504-8601-a08d7174edef)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" podUID="16eb2f20-a170-4504-8601-a08d7174edef" Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.895591 4874 scope.go:117] "RemoveContainer" containerID="09cb582dc76bac620de55013695a6049eb521a343c611714eaf77b7038c60d7e" Jan 22 12:06:08 crc kubenswrapper[4874]: I0122 12:06:08.982753 4874 scope.go:117] "RemoveContainer" containerID="d00586b99f40c23c83e98d796f708a92dd64728a1c6fc574bfa7124a09972b2f" Jan 22 12:06:09 crc kubenswrapper[4874]: I0122 12:06:09.877713 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq2h8" event={"ID":"be315a0a-e88a-4a39-b705-b826a4285fc2","Type":"ContainerStarted","Data":"198d6b87e2554ce51600165ca3a1b9773fd1588950afac105198296df391db5b"} Jan 22 12:06:09 crc kubenswrapper[4874]: I0122 12:06:09.879465 4874 generic.go:334] "Generic (PLEG): container finished" podID="1cca42d2-5390-4dc0-90fd-6abef972f828" containerID="03a5f9d616cc942c97a12adda95ac1fb8d538715f65c06b4d599618c1bca3dc2" exitCode=0 Jan 22 12:06:09 crc kubenswrapper[4874]: I0122 12:06:09.879552 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" event={"ID":"1cca42d2-5390-4dc0-90fd-6abef972f828","Type":"ContainerDied","Data":"03a5f9d616cc942c97a12adda95ac1fb8d538715f65c06b4d599618c1bca3dc2"} Jan 22 12:06:09 crc kubenswrapper[4874]: I0122 12:06:09.879618 4874 scope.go:117] "RemoveContainer" containerID="0b8c4a0b0ba7c449f8d752c598d619f6497f68d22e47bcf468a644699c35b9f8" Jan 22 12:06:09 crc kubenswrapper[4874]: I0122 12:06:09.880115 4874 scope.go:117] "RemoveContainer" containerID="03a5f9d616cc942c97a12adda95ac1fb8d538715f65c06b4d599618c1bca3dc2" Jan 22 12:06:09 crc kubenswrapper[4874]: E0122 12:06:09.880300 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np_service-telemetry(1cca42d2-5390-4dc0-90fd-6abef972f828)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" podUID="1cca42d2-5390-4dc0-90fd-6abef972f828" Jan 22 12:06:10 crc kubenswrapper[4874]: I0122 12:06:10.901998 4874 generic.go:334] "Generic (PLEG): container finished" podID="be315a0a-e88a-4a39-b705-b826a4285fc2" containerID="198d6b87e2554ce51600165ca3a1b9773fd1588950afac105198296df391db5b" exitCode=0 Jan 22 12:06:10 crc kubenswrapper[4874]: I0122 12:06:10.902107 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq2h8" event={"ID":"be315a0a-e88a-4a39-b705-b826a4285fc2","Type":"ContainerDied","Data":"198d6b87e2554ce51600165ca3a1b9773fd1588950afac105198296df391db5b"} Jan 22 12:06:13 crc kubenswrapper[4874]: I0122 12:06:13.520693 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:06:13 crc kubenswrapper[4874]: I0122 12:06:13.521070 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:06:18 crc kubenswrapper[4874]: I0122 12:06:18.958025 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"056ff179-9e03-4f48-b3d3-fa995d0c4093","Type":"ContainerStarted","Data":"c69f816f67c4796d2c865aeebd8564ed2b53358d7a60b573e98386afedcdfb4a"} Jan 22 12:06:18 crc kubenswrapper[4874]: I0122 12:06:18.960020 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq2h8" event={"ID":"be315a0a-e88a-4a39-b705-b826a4285fc2","Type":"ContainerStarted","Data":"d0a23445908893fbd592dcf2f022e7c2732381ff0f6eefd00b3e70840bac239c"} Jan 22 12:06:18 crc kubenswrapper[4874]: I0122 12:06:18.971042 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/qdr-test" podStartSLOduration=2.324265538 podStartE2EDuration="14.9710295s" podCreationTimestamp="2026-01-22 12:06:04 +0000 UTC" firstStartedPulling="2026-01-22 12:06:05.719533517 +0000 UTC m=+1539.564604587" lastFinishedPulling="2026-01-22 12:06:18.366297469 +0000 UTC m=+1552.211368549" observedRunningTime="2026-01-22 12:06:18.969840554 +0000 UTC m=+1552.814911624" watchObservedRunningTime="2026-01-22 12:06:18.9710295 +0000 UTC m=+1552.816100570" Jan 22 12:06:18 crc kubenswrapper[4874]: I0122 12:06:18.986618 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vq2h8" podStartSLOduration=3.426900384 podStartE2EDuration="12.986600998s" podCreationTimestamp="2026-01-22 12:06:06 +0000 UTC" firstStartedPulling="2026-01-22 12:06:08.791058979 +0000 UTC m=+1542.636130049" lastFinishedPulling="2026-01-22 12:06:18.350759593 +0000 UTC m=+1552.195830663" observedRunningTime="2026-01-22 12:06:18.984310347 +0000 UTC m=+1552.829381417" watchObservedRunningTime="2026-01-22 12:06:18.986600998 +0000 UTC m=+1552.831672058" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.653443 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-fpm2k"] Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.654458 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.656749 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-config" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.656782 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-healthcheck-log" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.657107 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-entrypoint-script" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.657110 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-entrypoint-script" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.657318 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-publisher" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.657668 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-sensubility-config" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.671842 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-fpm2k"] Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.715882 4874 scope.go:117] "RemoveContainer" containerID="241fe00fc340b80eefba0f92d3bdf67c1dc9e3306e497d8b33550b4e18947a08" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.749234 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-sensubility-config\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.749286 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-healthcheck-log\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.749352 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.749464 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-config\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.749508 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-publisher\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.749575 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqqp6\" (UniqueName: \"kubernetes.io/projected/91078bec-5716-4d9f-ade0-53de332b1871-kube-api-access-rqqp6\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.749602 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.850840 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-sensubility-config\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.850886 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-healthcheck-log\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.850941 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.851013 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-config\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.851050 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-publisher\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.851097 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqqp6\" (UniqueName: \"kubernetes.io/projected/91078bec-5716-4d9f-ade0-53de332b1871-kube-api-access-rqqp6\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.851114 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.852226 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-config\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.852735 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-publisher\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.853146 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-sensubility-config\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.853256 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-healthcheck-log\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.853501 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.853705 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.876290 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqqp6\" (UniqueName: \"kubernetes.io/projected/91078bec-5716-4d9f-ade0-53de332b1871-kube-api-access-rqqp6\") pod \"stf-smoketest-smoke1-fpm2k\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:19 crc kubenswrapper[4874]: I0122 12:06:19.973343 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.034859 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/curl"] Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.036004 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.069053 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.158004 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5wbx\" (UniqueName: \"kubernetes.io/projected/cd23c25e-9a15-4e31-be24-5d227c6ed353-kube-api-access-s5wbx\") pod \"curl\" (UID: \"cd23c25e-9a15-4e31-be24-5d227c6ed353\") " pod="service-telemetry/curl" Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.259703 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5wbx\" (UniqueName: \"kubernetes.io/projected/cd23c25e-9a15-4e31-be24-5d227c6ed353-kube-api-access-s5wbx\") pod \"curl\" (UID: \"cd23c25e-9a15-4e31-be24-5d227c6ed353\") " pod="service-telemetry/curl" Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.282201 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5wbx\" (UniqueName: \"kubernetes.io/projected/cd23c25e-9a15-4e31-be24-5d227c6ed353-kube-api-access-s5wbx\") pod \"curl\" (UID: \"cd23c25e-9a15-4e31-be24-5d227c6ed353\") " pod="service-telemetry/curl" Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.357524 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.412224 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-fpm2k"] Jan 22 12:06:20 crc kubenswrapper[4874]: W0122 12:06:20.420721 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91078bec_5716_4d9f_ade0_53de332b1871.slice/crio-16ed44bb34a808918bff425d12ba06fe0641976cb78ef34a60d35a441ffa9857 WatchSource:0}: Error finding container 16ed44bb34a808918bff425d12ba06fe0641976cb78ef34a60d35a441ffa9857: Status 404 returned error can't find the container with id 16ed44bb34a808918bff425d12ba06fe0641976cb78ef34a60d35a441ffa9857 Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.719500 4874 scope.go:117] "RemoveContainer" containerID="03a5f9d616cc942c97a12adda95ac1fb8d538715f65c06b4d599618c1bca3dc2" Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.720428 4874 scope.go:117] "RemoveContainer" containerID="9950cfa7893dfcae9ef6566159771b0e333f0fd846d8eadc4a654a109358eb87" Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.720805 4874 scope.go:117] "RemoveContainer" containerID="699a19bb9b06b8eb15d6f8ce1695ec22e8bc5543fd8a9c4aaee428742ce39ca8" Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.800441 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.980902 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7" event={"ID":"79746e6a-bee0-42eb-840c-36075d04abde","Type":"ContainerStarted","Data":"cea9253bbeb1e96e5fbe0807df172fd7931a2d7820e5965c098f7f75af5ec4d8"} Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.983640 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-fpm2k" event={"ID":"91078bec-5716-4d9f-ade0-53de332b1871","Type":"ContainerStarted","Data":"16ed44bb34a808918bff425d12ba06fe0641976cb78ef34a60d35a441ffa9857"} Jan 22 12:06:20 crc kubenswrapper[4874]: I0122 12:06:20.985106 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"cd23c25e-9a15-4e31-be24-5d227c6ed353","Type":"ContainerStarted","Data":"cd77744055f6f640939707b45933fd4765f68b786e5c80424cc767d9fde5ab83"} Jan 22 12:06:22 crc kubenswrapper[4874]: I0122 12:06:22.716480 4874 scope.go:117] "RemoveContainer" containerID="8871bcfaa155a6fc80df16ab23ea84232653fb97737ee28e90257e9bd644fd42" Jan 22 12:06:23 crc kubenswrapper[4874]: I0122 12:06:23.018057 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np" event={"ID":"1cca42d2-5390-4dc0-90fd-6abef972f828","Type":"ContainerStarted","Data":"0690e473692dea624820ef6cdbdca8f8bad3bda1058f3a6c42e0cbf6b26d1e8e"} Jan 22 12:06:23 crc kubenswrapper[4874]: I0122 12:06:23.056379 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg" event={"ID":"16eb2f20-a170-4504-8601-a08d7174edef","Type":"ContainerStarted","Data":"cfd3081857be454b30191ece8c21d3207394a76215241e108f66bf635fa62517"} Jan 22 12:06:24 crc kubenswrapper[4874]: I0122 12:06:24.067367 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf" event={"ID":"7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2","Type":"ContainerStarted","Data":"b368544b902c3c3ab9b296ade349185e174c7a61d1ba17c7faa0ef3f89219855"} Jan 22 12:06:24 crc kubenswrapper[4874]: I0122 12:06:24.070489 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9" event={"ID":"01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3","Type":"ContainerStarted","Data":"ac263c6d72ebfd50563b514b8917992ee1d7623d1fe32c5985f30e4b885f2b64"} Jan 22 12:06:25 crc kubenswrapper[4874]: I0122 12:06:25.090158 4874 generic.go:334] "Generic (PLEG): container finished" podID="cd23c25e-9a15-4e31-be24-5d227c6ed353" containerID="3632d001d41bcbca464db121b0d9020680ba26e3f8c8d7e16a16e237f6ec3f69" exitCode=0 Jan 22 12:06:25 crc kubenswrapper[4874]: I0122 12:06:25.090251 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"cd23c25e-9a15-4e31-be24-5d227c6ed353","Type":"ContainerDied","Data":"3632d001d41bcbca464db121b0d9020680ba26e3f8c8d7e16a16e237f6ec3f69"} Jan 22 12:06:27 crc kubenswrapper[4874]: I0122 12:06:27.225127 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:27 crc kubenswrapper[4874]: I0122 12:06:27.225498 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:27 crc kubenswrapper[4874]: I0122 12:06:27.267877 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:28 crc kubenswrapper[4874]: I0122 12:06:28.158052 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:28 crc kubenswrapper[4874]: I0122 12:06:28.205700 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vq2h8"] Jan 22 12:06:28 crc kubenswrapper[4874]: I0122 12:06:28.622143 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 22 12:06:28 crc kubenswrapper[4874]: I0122 12:06:28.702594 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5wbx\" (UniqueName: \"kubernetes.io/projected/cd23c25e-9a15-4e31-be24-5d227c6ed353-kube-api-access-s5wbx\") pod \"cd23c25e-9a15-4e31-be24-5d227c6ed353\" (UID: \"cd23c25e-9a15-4e31-be24-5d227c6ed353\") " Jan 22 12:06:28 crc kubenswrapper[4874]: I0122 12:06:28.721009 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd23c25e-9a15-4e31-be24-5d227c6ed353-kube-api-access-s5wbx" (OuterVolumeSpecName: "kube-api-access-s5wbx") pod "cd23c25e-9a15-4e31-be24-5d227c6ed353" (UID: "cd23c25e-9a15-4e31-be24-5d227c6ed353"). InnerVolumeSpecName "kube-api-access-s5wbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:06:28 crc kubenswrapper[4874]: I0122 12:06:28.781839 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_curl_cd23c25e-9a15-4e31-be24-5d227c6ed353/curl/0.log" Jan 22 12:06:28 crc kubenswrapper[4874]: I0122 12:06:28.804539 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5wbx\" (UniqueName: \"kubernetes.io/projected/cd23c25e-9a15-4e31-be24-5d227c6ed353-kube-api-access-s5wbx\") on node \"crc\" DevicePath \"\"" Jan 22 12:06:29 crc kubenswrapper[4874]: I0122 12:06:29.112011 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-9qcrx_a14f7a5d-1819-4ca2-8683-96a338c70df6/prometheus-webhook-snmp/0.log" Jan 22 12:06:29 crc kubenswrapper[4874]: I0122 12:06:29.121733 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"cd23c25e-9a15-4e31-be24-5d227c6ed353","Type":"ContainerDied","Data":"cd77744055f6f640939707b45933fd4765f68b786e5c80424cc767d9fde5ab83"} Jan 22 12:06:29 crc kubenswrapper[4874]: I0122 12:06:29.121769 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd77744055f6f640939707b45933fd4765f68b786e5c80424cc767d9fde5ab83" Jan 22 12:06:29 crc kubenswrapper[4874]: I0122 12:06:29.121872 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 22 12:06:30 crc kubenswrapper[4874]: I0122 12:06:30.129717 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vq2h8" podUID="be315a0a-e88a-4a39-b705-b826a4285fc2" containerName="registry-server" containerID="cri-o://d0a23445908893fbd592dcf2f022e7c2732381ff0f6eefd00b3e70840bac239c" gracePeriod=2 Jan 22 12:06:31 crc kubenswrapper[4874]: I0122 12:06:31.139944 4874 generic.go:334] "Generic (PLEG): container finished" podID="be315a0a-e88a-4a39-b705-b826a4285fc2" containerID="d0a23445908893fbd592dcf2f022e7c2732381ff0f6eefd00b3e70840bac239c" exitCode=0 Jan 22 12:06:31 crc kubenswrapper[4874]: I0122 12:06:31.139985 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq2h8" event={"ID":"be315a0a-e88a-4a39-b705-b826a4285fc2","Type":"ContainerDied","Data":"d0a23445908893fbd592dcf2f022e7c2732381ff0f6eefd00b3e70840bac239c"} Jan 22 12:06:33 crc kubenswrapper[4874]: I0122 12:06:33.415474 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:33 crc kubenswrapper[4874]: I0122 12:06:33.569956 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj82v\" (UniqueName: \"kubernetes.io/projected/be315a0a-e88a-4a39-b705-b826a4285fc2-kube-api-access-cj82v\") pod \"be315a0a-e88a-4a39-b705-b826a4285fc2\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " Jan 22 12:06:33 crc kubenswrapper[4874]: I0122 12:06:33.570150 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-utilities\") pod \"be315a0a-e88a-4a39-b705-b826a4285fc2\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " Jan 22 12:06:33 crc kubenswrapper[4874]: I0122 12:06:33.570235 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-catalog-content\") pod \"be315a0a-e88a-4a39-b705-b826a4285fc2\" (UID: \"be315a0a-e88a-4a39-b705-b826a4285fc2\") " Jan 22 12:06:33 crc kubenswrapper[4874]: I0122 12:06:33.571130 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-utilities" (OuterVolumeSpecName: "utilities") pod "be315a0a-e88a-4a39-b705-b826a4285fc2" (UID: "be315a0a-e88a-4a39-b705-b826a4285fc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:06:33 crc kubenswrapper[4874]: I0122 12:06:33.587072 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be315a0a-e88a-4a39-b705-b826a4285fc2-kube-api-access-cj82v" (OuterVolumeSpecName: "kube-api-access-cj82v") pod "be315a0a-e88a-4a39-b705-b826a4285fc2" (UID: "be315a0a-e88a-4a39-b705-b826a4285fc2"). InnerVolumeSpecName "kube-api-access-cj82v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:06:33 crc kubenswrapper[4874]: I0122 12:06:33.648917 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be315a0a-e88a-4a39-b705-b826a4285fc2" (UID: "be315a0a-e88a-4a39-b705-b826a4285fc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:06:33 crc kubenswrapper[4874]: I0122 12:06:33.671668 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:06:33 crc kubenswrapper[4874]: I0122 12:06:33.671698 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be315a0a-e88a-4a39-b705-b826a4285fc2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:06:33 crc kubenswrapper[4874]: I0122 12:06:33.671712 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj82v\" (UniqueName: \"kubernetes.io/projected/be315a0a-e88a-4a39-b705-b826a4285fc2-kube-api-access-cj82v\") on node \"crc\" DevicePath \"\"" Jan 22 12:06:34 crc kubenswrapper[4874]: I0122 12:06:34.167103 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq2h8" event={"ID":"be315a0a-e88a-4a39-b705-b826a4285fc2","Type":"ContainerDied","Data":"6095ec06d7054068d7a98b6e4ccf292ac930e508cf3c276df0c4600afa1fd83e"} Jan 22 12:06:34 crc kubenswrapper[4874]: I0122 12:06:34.167187 4874 scope.go:117] "RemoveContainer" containerID="d0a23445908893fbd592dcf2f022e7c2732381ff0f6eefd00b3e70840bac239c" Jan 22 12:06:34 crc kubenswrapper[4874]: I0122 12:06:34.167134 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vq2h8" Jan 22 12:06:34 crc kubenswrapper[4874]: I0122 12:06:34.168898 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-fpm2k" event={"ID":"91078bec-5716-4d9f-ade0-53de332b1871","Type":"ContainerStarted","Data":"38b91d3438b9959ba5ddfc6ef9264d72a0bd2d404dd8c5472433d88d1af78849"} Jan 22 12:06:34 crc kubenswrapper[4874]: I0122 12:06:34.198773 4874 scope.go:117] "RemoveContainer" containerID="198d6b87e2554ce51600165ca3a1b9773fd1588950afac105198296df391db5b" Jan 22 12:06:34 crc kubenswrapper[4874]: I0122 12:06:34.222359 4874 scope.go:117] "RemoveContainer" containerID="4119410dea713ff391706990d3ca3df3ffd65b7a0e739319efe9d22e1aefc65b" Jan 22 12:06:34 crc kubenswrapper[4874]: I0122 12:06:34.223670 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vq2h8"] Jan 22 12:06:34 crc kubenswrapper[4874]: I0122 12:06:34.231809 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vq2h8"] Jan 22 12:06:34 crc kubenswrapper[4874]: I0122 12:06:34.733247 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be315a0a-e88a-4a39-b705-b826a4285fc2" path="/var/lib/kubelet/pods/be315a0a-e88a-4a39-b705-b826a4285fc2/volumes" Jan 22 12:06:39 crc kubenswrapper[4874]: I0122 12:06:39.211452 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-fpm2k" event={"ID":"91078bec-5716-4d9f-ade0-53de332b1871","Type":"ContainerStarted","Data":"19047b524cd4ee68868d7fda810d116221c6871b036d5650ee13bfaeb07ec127"} Jan 22 12:06:39 crc kubenswrapper[4874]: I0122 12:06:39.245034 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-fpm2k" podStartSLOduration=1.98738127 podStartE2EDuration="20.245007741s" podCreationTimestamp="2026-01-22 12:06:19 +0000 UTC" firstStartedPulling="2026-01-22 12:06:20.422426569 +0000 UTC m=+1554.267497639" lastFinishedPulling="2026-01-22 12:06:38.68005304 +0000 UTC m=+1572.525124110" observedRunningTime="2026-01-22 12:06:39.234474014 +0000 UTC m=+1573.079545174" watchObservedRunningTime="2026-01-22 12:06:39.245007741 +0000 UTC m=+1573.090078851" Jan 22 12:06:43 crc kubenswrapper[4874]: I0122 12:06:43.520474 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:06:43 crc kubenswrapper[4874]: I0122 12:06:43.520835 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:06:43 crc kubenswrapper[4874]: I0122 12:06:43.520881 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 12:06:43 crc kubenswrapper[4874]: I0122 12:06:43.521527 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 12:06:43 crc kubenswrapper[4874]: I0122 12:06:43.521585 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" gracePeriod=600 Jan 22 12:06:43 crc kubenswrapper[4874]: E0122 12:06:43.657483 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:06:44 crc kubenswrapper[4874]: I0122 12:06:44.260245 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" exitCode=0 Jan 22 12:06:44 crc kubenswrapper[4874]: I0122 12:06:44.260336 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15"} Jan 22 12:06:44 crc kubenswrapper[4874]: I0122 12:06:44.260445 4874 scope.go:117] "RemoveContainer" containerID="75a7548055039130cae1c6ba4be8efe015a8bd6db752337a976c1be606300781" Jan 22 12:06:44 crc kubenswrapper[4874]: I0122 12:06:44.261263 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:06:44 crc kubenswrapper[4874]: E0122 12:06:44.261996 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:06:58 crc kubenswrapper[4874]: I0122 12:06:58.717092 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:06:58 crc kubenswrapper[4874]: E0122 12:06:58.718209 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:06:59 crc kubenswrapper[4874]: I0122 12:06:59.275083 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-9qcrx_a14f7a5d-1819-4ca2-8683-96a338c70df6/prometheus-webhook-snmp/0.log" Jan 22 12:07:07 crc kubenswrapper[4874]: I0122 12:07:07.460641 4874 generic.go:334] "Generic (PLEG): container finished" podID="91078bec-5716-4d9f-ade0-53de332b1871" containerID="38b91d3438b9959ba5ddfc6ef9264d72a0bd2d404dd8c5472433d88d1af78849" exitCode=1 Jan 22 12:07:07 crc kubenswrapper[4874]: I0122 12:07:07.460736 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-fpm2k" event={"ID":"91078bec-5716-4d9f-ade0-53de332b1871","Type":"ContainerDied","Data":"38b91d3438b9959ba5ddfc6ef9264d72a0bd2d404dd8c5472433d88d1af78849"} Jan 22 12:07:07 crc kubenswrapper[4874]: I0122 12:07:07.462500 4874 scope.go:117] "RemoveContainer" containerID="38b91d3438b9959ba5ddfc6ef9264d72a0bd2d404dd8c5472433d88d1af78849" Jan 22 12:07:10 crc kubenswrapper[4874]: I0122 12:07:10.491960 4874 generic.go:334] "Generic (PLEG): container finished" podID="91078bec-5716-4d9f-ade0-53de332b1871" containerID="19047b524cd4ee68868d7fda810d116221c6871b036d5650ee13bfaeb07ec127" exitCode=0 Jan 22 12:07:10 crc kubenswrapper[4874]: I0122 12:07:10.492616 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-fpm2k" event={"ID":"91078bec-5716-4d9f-ade0-53de332b1871","Type":"ContainerDied","Data":"19047b524cd4ee68868d7fda810d116221c6871b036d5650ee13bfaeb07ec127"} Jan 22 12:07:10 crc kubenswrapper[4874]: I0122 12:07:10.718420 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:07:10 crc kubenswrapper[4874]: E0122 12:07:10.719184 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:07:11 crc kubenswrapper[4874]: I0122 12:07:11.881303 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:07:11 crc kubenswrapper[4874]: I0122 12:07:11.991889 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-sensubility-config\") pod \"91078bec-5716-4d9f-ade0-53de332b1871\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " Jan 22 12:07:11 crc kubenswrapper[4874]: I0122 12:07:11.991926 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-publisher\") pod \"91078bec-5716-4d9f-ade0-53de332b1871\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " Jan 22 12:07:11 crc kubenswrapper[4874]: I0122 12:07:11.992021 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqqp6\" (UniqueName: \"kubernetes.io/projected/91078bec-5716-4d9f-ade0-53de332b1871-kube-api-access-rqqp6\") pod \"91078bec-5716-4d9f-ade0-53de332b1871\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " Jan 22 12:07:11 crc kubenswrapper[4874]: I0122 12:07:11.992049 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-healthcheck-log\") pod \"91078bec-5716-4d9f-ade0-53de332b1871\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " Jan 22 12:07:11 crc kubenswrapper[4874]: I0122 12:07:11.992076 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-entrypoint-script\") pod \"91078bec-5716-4d9f-ade0-53de332b1871\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " Jan 22 12:07:11 crc kubenswrapper[4874]: I0122 12:07:11.992105 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-config\") pod \"91078bec-5716-4d9f-ade0-53de332b1871\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " Jan 22 12:07:11 crc kubenswrapper[4874]: I0122 12:07:11.992150 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-entrypoint-script\") pod \"91078bec-5716-4d9f-ade0-53de332b1871\" (UID: \"91078bec-5716-4d9f-ade0-53de332b1871\") " Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.000128 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91078bec-5716-4d9f-ade0-53de332b1871-kube-api-access-rqqp6" (OuterVolumeSpecName: "kube-api-access-rqqp6") pod "91078bec-5716-4d9f-ade0-53de332b1871" (UID: "91078bec-5716-4d9f-ade0-53de332b1871"). InnerVolumeSpecName "kube-api-access-rqqp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.011081 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "91078bec-5716-4d9f-ade0-53de332b1871" (UID: "91078bec-5716-4d9f-ade0-53de332b1871"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.012294 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "91078bec-5716-4d9f-ade0-53de332b1871" (UID: "91078bec-5716-4d9f-ade0-53de332b1871"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.012845 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "91078bec-5716-4d9f-ade0-53de332b1871" (UID: "91078bec-5716-4d9f-ade0-53de332b1871"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.014369 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "91078bec-5716-4d9f-ade0-53de332b1871" (UID: "91078bec-5716-4d9f-ade0-53de332b1871"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.015990 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "91078bec-5716-4d9f-ade0-53de332b1871" (UID: "91078bec-5716-4d9f-ade0-53de332b1871"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.023021 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "91078bec-5716-4d9f-ade0-53de332b1871" (UID: "91078bec-5716-4d9f-ade0-53de332b1871"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.093796 4874 reconciler_common.go:293] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-sensubility-config\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.093893 4874 reconciler_common.go:293] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.093909 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqqp6\" (UniqueName: \"kubernetes.io/projected/91078bec-5716-4d9f-ade0-53de332b1871-kube-api-access-rqqp6\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.093918 4874 reconciler_common.go:293] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-healthcheck-log\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.093929 4874 reconciler_common.go:293] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.093939 4874 reconciler_common.go:293] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-collectd-config\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.093947 4874 reconciler_common.go:293] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/91078bec-5716-4d9f-ade0-53de332b1871-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.513022 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-fpm2k" event={"ID":"91078bec-5716-4d9f-ade0-53de332b1871","Type":"ContainerDied","Data":"16ed44bb34a808918bff425d12ba06fe0641976cb78ef34a60d35a441ffa9857"} Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.513079 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16ed44bb34a808918bff425d12ba06fe0641976cb78ef34a60d35a441ffa9857" Jan 22 12:07:12 crc kubenswrapper[4874]: I0122 12:07:12.513100 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-fpm2k" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.042687 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-qb5kr"] Jan 22 12:07:20 crc kubenswrapper[4874]: E0122 12:07:20.044054 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91078bec-5716-4d9f-ade0-53de332b1871" containerName="smoketest-ceilometer" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.044087 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="91078bec-5716-4d9f-ade0-53de332b1871" containerName="smoketest-ceilometer" Jan 22 12:07:20 crc kubenswrapper[4874]: E0122 12:07:20.044114 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be315a0a-e88a-4a39-b705-b826a4285fc2" containerName="extract-utilities" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.044131 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="be315a0a-e88a-4a39-b705-b826a4285fc2" containerName="extract-utilities" Jan 22 12:07:20 crc kubenswrapper[4874]: E0122 12:07:20.044147 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91078bec-5716-4d9f-ade0-53de332b1871" containerName="smoketest-collectd" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.044169 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="91078bec-5716-4d9f-ade0-53de332b1871" containerName="smoketest-collectd" Jan 22 12:07:20 crc kubenswrapper[4874]: E0122 12:07:20.044190 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd23c25e-9a15-4e31-be24-5d227c6ed353" containerName="curl" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.044206 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd23c25e-9a15-4e31-be24-5d227c6ed353" containerName="curl" Jan 22 12:07:20 crc kubenswrapper[4874]: E0122 12:07:20.044240 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be315a0a-e88a-4a39-b705-b826a4285fc2" containerName="extract-content" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.044255 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="be315a0a-e88a-4a39-b705-b826a4285fc2" containerName="extract-content" Jan 22 12:07:20 crc kubenswrapper[4874]: E0122 12:07:20.044288 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be315a0a-e88a-4a39-b705-b826a4285fc2" containerName="registry-server" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.044300 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="be315a0a-e88a-4a39-b705-b826a4285fc2" containerName="registry-server" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.044539 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd23c25e-9a15-4e31-be24-5d227c6ed353" containerName="curl" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.044557 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="91078bec-5716-4d9f-ade0-53de332b1871" containerName="smoketest-ceilometer" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.044582 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="be315a0a-e88a-4a39-b705-b826a4285fc2" containerName="registry-server" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.044600 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="91078bec-5716-4d9f-ade0-53de332b1871" containerName="smoketest-collectd" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.046130 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.049800 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-sensubility-config" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.050517 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-entrypoint-script" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.051034 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-qb5kr"] Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.051672 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-entrypoint-script" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.052126 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-healthcheck-log" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.052303 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-publisher" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.052836 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-config" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.129294 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mt5z\" (UniqueName: \"kubernetes.io/projected/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-kube-api-access-2mt5z\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.129347 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.129468 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-publisher\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.129527 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-config\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.129558 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-sensubility-config\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.129597 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.129642 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-healthcheck-log\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.231744 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-publisher\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.231956 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-config\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.232148 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-sensubility-config\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.232345 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.232465 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-healthcheck-log\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.232588 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mt5z\" (UniqueName: \"kubernetes.io/projected/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-kube-api-access-2mt5z\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.232741 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.233962 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-sensubility-config\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.234037 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.234110 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-config\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.234427 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.235186 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-healthcheck-log\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.235714 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-publisher\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.258256 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mt5z\" (UniqueName: \"kubernetes.io/projected/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-kube-api-access-2mt5z\") pod \"stf-smoketest-smoke1-qb5kr\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.394888 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:20 crc kubenswrapper[4874]: I0122 12:07:20.667773 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-qb5kr"] Jan 22 12:07:21 crc kubenswrapper[4874]: I0122 12:07:21.602243 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-qb5kr" event={"ID":"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade","Type":"ContainerStarted","Data":"5db0d7d6f5affaaef96d0acc5b44b5c8ab6420c145f93bedc894a1712c50851b"} Jan 22 12:07:21 crc kubenswrapper[4874]: I0122 12:07:21.602952 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-qb5kr" event={"ID":"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade","Type":"ContainerStarted","Data":"1cdaa03af8fbf3738ed2d9e7fe71986bd4c0931da9f26d90d6445d92aef346c4"} Jan 22 12:07:21 crc kubenswrapper[4874]: I0122 12:07:21.602980 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-qb5kr" event={"ID":"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade","Type":"ContainerStarted","Data":"3c6c3534d1ad955c051613aa5e3cd4f190eb991eaeb3ea432b665596e6886c77"} Jan 22 12:07:21 crc kubenswrapper[4874]: I0122 12:07:21.632067 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-qb5kr" podStartSLOduration=1.6320396019999999 podStartE2EDuration="1.632039602s" podCreationTimestamp="2026-01-22 12:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 12:07:21.629247695 +0000 UTC m=+1615.474318765" watchObservedRunningTime="2026-01-22 12:07:21.632039602 +0000 UTC m=+1615.477110672" Jan 22 12:07:23 crc kubenswrapper[4874]: I0122 12:07:23.716365 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:07:23 crc kubenswrapper[4874]: E0122 12:07:23.717111 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:07:35 crc kubenswrapper[4874]: I0122 12:07:35.715895 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:07:35 crc kubenswrapper[4874]: E0122 12:07:35.717459 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:07:49 crc kubenswrapper[4874]: I0122 12:07:49.715975 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:07:49 crc kubenswrapper[4874]: E0122 12:07:49.717039 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:07:53 crc kubenswrapper[4874]: I0122 12:07:53.886939 4874 generic.go:334] "Generic (PLEG): container finished" podID="c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" containerID="5db0d7d6f5affaaef96d0acc5b44b5c8ab6420c145f93bedc894a1712c50851b" exitCode=0 Jan 22 12:07:53 crc kubenswrapper[4874]: I0122 12:07:53.887015 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-qb5kr" event={"ID":"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade","Type":"ContainerDied","Data":"5db0d7d6f5affaaef96d0acc5b44b5c8ab6420c145f93bedc894a1712c50851b"} Jan 22 12:07:53 crc kubenswrapper[4874]: I0122 12:07:53.888502 4874 scope.go:117] "RemoveContainer" containerID="5db0d7d6f5affaaef96d0acc5b44b5c8ab6420c145f93bedc894a1712c50851b" Jan 22 12:07:54 crc kubenswrapper[4874]: I0122 12:07:54.896642 4874 generic.go:334] "Generic (PLEG): container finished" podID="c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" containerID="1cdaa03af8fbf3738ed2d9e7fe71986bd4c0931da9f26d90d6445d92aef346c4" exitCode=0 Jan 22 12:07:54 crc kubenswrapper[4874]: I0122 12:07:54.896693 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-qb5kr" event={"ID":"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade","Type":"ContainerDied","Data":"1cdaa03af8fbf3738ed2d9e7fe71986bd4c0931da9f26d90d6445d92aef346c4"} Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.290205 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.486696 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-sensubility-config\") pod \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.487045 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-entrypoint-script\") pod \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.487229 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-publisher\") pod \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.487363 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-healthcheck-log\") pod \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.487539 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mt5z\" (UniqueName: \"kubernetes.io/projected/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-kube-api-access-2mt5z\") pod \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.487730 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-config\") pod \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.487863 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-entrypoint-script\") pod \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\" (UID: \"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade\") " Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.496738 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-kube-api-access-2mt5z" (OuterVolumeSpecName: "kube-api-access-2mt5z") pod "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" (UID: "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade"). InnerVolumeSpecName "kube-api-access-2mt5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.506772 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" (UID: "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.509471 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" (UID: "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.510200 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" (UID: "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.512604 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" (UID: "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.515133 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" (UID: "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.529096 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" (UID: "c1f3f1ed-6e23-4ad5-877c-522b9cb42ade"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.589371 4874 reconciler_common.go:293] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.589461 4874 reconciler_common.go:293] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.589480 4874 reconciler_common.go:293] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-healthcheck-log\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.589500 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mt5z\" (UniqueName: \"kubernetes.io/projected/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-kube-api-access-2mt5z\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.589517 4874 reconciler_common.go:293] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-collectd-config\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.589533 4874 reconciler_common.go:293] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.589549 4874 reconciler_common.go:293] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/c1f3f1ed-6e23-4ad5-877c-522b9cb42ade-sensubility-config\") on node \"crc\" DevicePath \"\"" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.917620 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-qb5kr" event={"ID":"c1f3f1ed-6e23-4ad5-877c-522b9cb42ade","Type":"ContainerDied","Data":"3c6c3534d1ad955c051613aa5e3cd4f190eb991eaeb3ea432b665596e6886c77"} Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.917715 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c6c3534d1ad955c051613aa5e3cd4f190eb991eaeb3ea432b665596e6886c77" Jan 22 12:07:56 crc kubenswrapper[4874]: I0122 12:07:56.917800 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-qb5kr" Jan 22 12:07:58 crc kubenswrapper[4874]: I0122 12:07:58.510299 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-fpm2k_91078bec-5716-4d9f-ade0-53de332b1871/smoketest-collectd/0.log" Jan 22 12:07:58 crc kubenswrapper[4874]: I0122 12:07:58.869748 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-fpm2k_91078bec-5716-4d9f-ade0-53de332b1871/smoketest-ceilometer/0.log" Jan 22 12:07:59 crc kubenswrapper[4874]: I0122 12:07:59.162535 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-interconnect-68864d46cb-jcvt5_da6dd65e-8b70-41a1-8a38-83392c65aef3/default-interconnect/0.log" Jan 22 12:07:59 crc kubenswrapper[4874]: I0122 12:07:59.479110 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg_16eb2f20-a170-4504-8601-a08d7174edef/bridge/2.log" Jan 22 12:07:59 crc kubenswrapper[4874]: I0122 12:07:59.764628 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg_16eb2f20-a170-4504-8601-a08d7174edef/sg-core/0.log" Jan 22 12:08:00 crc kubenswrapper[4874]: I0122 12:08:00.013565 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7_79746e6a-bee0-42eb-840c-36075d04abde/bridge/2.log" Jan 22 12:08:00 crc kubenswrapper[4874]: I0122 12:08:00.296986 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7_79746e6a-bee0-42eb-840c-36075d04abde/sg-core/0.log" Jan 22 12:08:00 crc kubenswrapper[4874]: I0122 12:08:00.586517 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np_1cca42d2-5390-4dc0-90fd-6abef972f828/bridge/2.log" Jan 22 12:08:00 crc kubenswrapper[4874]: I0122 12:08:00.846144 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np_1cca42d2-5390-4dc0-90fd-6abef972f828/sg-core/0.log" Jan 22 12:08:01 crc kubenswrapper[4874]: I0122 12:08:01.111598 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9_01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3/bridge/2.log" Jan 22 12:08:01 crc kubenswrapper[4874]: I0122 12:08:01.364928 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9_01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3/sg-core/0.log" Jan 22 12:08:01 crc kubenswrapper[4874]: I0122 12:08:01.614024 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf_7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2/bridge/2.log" Jan 22 12:08:01 crc kubenswrapper[4874]: I0122 12:08:01.857007 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf_7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2/sg-core/0.log" Jan 22 12:08:03 crc kubenswrapper[4874]: I0122 12:08:03.716258 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:08:03 crc kubenswrapper[4874]: E0122 12:08:03.716970 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:08:05 crc kubenswrapper[4874]: I0122 12:08:05.681337 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-779b7b47dc-4zkhw_f01ecec4-43d6-424e-843b-ef89a732199e/operator/0.log" Jan 22 12:08:06 crc kubenswrapper[4874]: I0122 12:08:06.959776 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_7cd2b3c5-32a6-49c1-984c-86aa8fe36f85/prometheus/0.log" Jan 22 12:08:07 crc kubenswrapper[4874]: I0122 12:08:07.235322 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_c9fa7f48-985a-46b3-a7dc-1b39dfc14243/elasticsearch/0.log" Jan 22 12:08:07 crc kubenswrapper[4874]: I0122 12:08:07.498051 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-9qcrx_a14f7a5d-1819-4ca2-8683-96a338c70df6/prometheus-webhook-snmp/0.log" Jan 22 12:08:07 crc kubenswrapper[4874]: I0122 12:08:07.790178 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_30722d93-5804-4aba-a6e3-e2b891356163/alertmanager/0.log" Jan 22 12:08:14 crc kubenswrapper[4874]: I0122 12:08:14.717178 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:08:14 crc kubenswrapper[4874]: E0122 12:08:14.717991 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:08:23 crc kubenswrapper[4874]: I0122 12:08:23.055527 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-66768bfd46-6v9td_9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413/operator/0.log" Jan 22 12:08:26 crc kubenswrapper[4874]: I0122 12:08:26.578286 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-779b7b47dc-4zkhw_f01ecec4-43d6-424e-843b-ef89a732199e/operator/0.log" Jan 22 12:08:26 crc kubenswrapper[4874]: I0122 12:08:26.914077 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_qdr-test_056ff179-9e03-4f48-b3d3-fa995d0c4093/qdr/0.log" Jan 22 12:08:27 crc kubenswrapper[4874]: I0122 12:08:27.716823 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:08:27 crc kubenswrapper[4874]: E0122 12:08:27.717212 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:08:39 crc kubenswrapper[4874]: I0122 12:08:39.716819 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:08:39 crc kubenswrapper[4874]: E0122 12:08:39.720306 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.392065 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-p8jdn/must-gather-vjgnn"] Jan 22 12:08:51 crc kubenswrapper[4874]: E0122 12:08:51.392737 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" containerName="smoketest-collectd" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.392750 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" containerName="smoketest-collectd" Jan 22 12:08:51 crc kubenswrapper[4874]: E0122 12:08:51.392759 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" containerName="smoketest-ceilometer" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.392765 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" containerName="smoketest-ceilometer" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.392881 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" containerName="smoketest-collectd" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.392902 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1f3f1ed-6e23-4ad5-877c-522b9cb42ade" containerName="smoketest-ceilometer" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.393631 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p8jdn/must-gather-vjgnn" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.396412 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-p8jdn"/"kube-root-ca.crt" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.396607 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-p8jdn"/"default-dockercfg-fcw2b" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.396619 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-p8jdn"/"openshift-service-ca.crt" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.406856 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-p8jdn/must-gather-vjgnn"] Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.509609 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6lg5\" (UniqueName: \"kubernetes.io/projected/c0158516-a991-4b44-a001-e94e467d86bb-kube-api-access-c6lg5\") pod \"must-gather-vjgnn\" (UID: \"c0158516-a991-4b44-a001-e94e467d86bb\") " pod="openshift-must-gather-p8jdn/must-gather-vjgnn" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.509694 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c0158516-a991-4b44-a001-e94e467d86bb-must-gather-output\") pod \"must-gather-vjgnn\" (UID: \"c0158516-a991-4b44-a001-e94e467d86bb\") " pod="openshift-must-gather-p8jdn/must-gather-vjgnn" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.611365 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c0158516-a991-4b44-a001-e94e467d86bb-must-gather-output\") pod \"must-gather-vjgnn\" (UID: \"c0158516-a991-4b44-a001-e94e467d86bb\") " pod="openshift-must-gather-p8jdn/must-gather-vjgnn" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.611473 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6lg5\" (UniqueName: \"kubernetes.io/projected/c0158516-a991-4b44-a001-e94e467d86bb-kube-api-access-c6lg5\") pod \"must-gather-vjgnn\" (UID: \"c0158516-a991-4b44-a001-e94e467d86bb\") " pod="openshift-must-gather-p8jdn/must-gather-vjgnn" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.611840 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c0158516-a991-4b44-a001-e94e467d86bb-must-gather-output\") pod \"must-gather-vjgnn\" (UID: \"c0158516-a991-4b44-a001-e94e467d86bb\") " pod="openshift-must-gather-p8jdn/must-gather-vjgnn" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.646039 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6lg5\" (UniqueName: \"kubernetes.io/projected/c0158516-a991-4b44-a001-e94e467d86bb-kube-api-access-c6lg5\") pod \"must-gather-vjgnn\" (UID: \"c0158516-a991-4b44-a001-e94e467d86bb\") " pod="openshift-must-gather-p8jdn/must-gather-vjgnn" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.719248 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p8jdn/must-gather-vjgnn" Jan 22 12:08:51 crc kubenswrapper[4874]: I0122 12:08:51.926136 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-p8jdn/must-gather-vjgnn"] Jan 22 12:08:52 crc kubenswrapper[4874]: I0122 12:08:52.431064 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p8jdn/must-gather-vjgnn" event={"ID":"c0158516-a991-4b44-a001-e94e467d86bb","Type":"ContainerStarted","Data":"c77ee96dabe444d06835fd65909ea3e64c10ef2ade708b20a2ec3f63f8d90fd9"} Jan 22 12:08:54 crc kubenswrapper[4874]: I0122 12:08:54.716566 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:08:54 crc kubenswrapper[4874]: E0122 12:08:54.716998 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:08:59 crc kubenswrapper[4874]: I0122 12:08:59.492683 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p8jdn/must-gather-vjgnn" event={"ID":"c0158516-a991-4b44-a001-e94e467d86bb","Type":"ContainerStarted","Data":"4f99b8d22b3c6ac4771f24a766d4e5db8e117f232ae19fdaee5bbeae454a8a40"} Jan 22 12:08:59 crc kubenswrapper[4874]: I0122 12:08:59.494451 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p8jdn/must-gather-vjgnn" event={"ID":"c0158516-a991-4b44-a001-e94e467d86bb","Type":"ContainerStarted","Data":"7b79b2c0b847e1079a6673f2439895793e3cb542df4be3a5ec6e2d56fc4f8a88"} Jan 22 12:08:59 crc kubenswrapper[4874]: I0122 12:08:59.516105 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-p8jdn/must-gather-vjgnn" podStartSLOduration=1.4874119860000001 podStartE2EDuration="8.516078843s" podCreationTimestamp="2026-01-22 12:08:51 +0000 UTC" firstStartedPulling="2026-01-22 12:08:51.939875555 +0000 UTC m=+1705.784946625" lastFinishedPulling="2026-01-22 12:08:58.968542402 +0000 UTC m=+1712.813613482" observedRunningTime="2026-01-22 12:08:59.50660852 +0000 UTC m=+1713.351679600" watchObservedRunningTime="2026-01-22 12:08:59.516078843 +0000 UTC m=+1713.361149933" Jan 22 12:09:09 crc kubenswrapper[4874]: I0122 12:09:09.716513 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:09:09 crc kubenswrapper[4874]: E0122 12:09:09.717300 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:09:13 crc kubenswrapper[4874]: I0122 12:09:13.094381 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6jqjk_a785e748-a557-4c24-8a4d-dc03cfc3c357/control-plane-machine-set-operator/0.log" Jan 22 12:09:13 crc kubenswrapper[4874]: I0122 12:09:13.122393 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wmwcz_9bec261f-fd1e-44f7-a402-cae34f722b6c/kube-rbac-proxy/0.log" Jan 22 12:09:13 crc kubenswrapper[4874]: I0122 12:09:13.133631 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wmwcz_9bec261f-fd1e-44f7-a402-cae34f722b6c/machine-api-operator/0.log" Jan 22 12:09:18 crc kubenswrapper[4874]: I0122 12:09:18.447073 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-7nhkf_50438a7c-3a86-4aeb-a9c6-10ccb68f4593/cert-manager-controller/0.log" Jan 22 12:09:18 crc kubenswrapper[4874]: I0122 12:09:18.465813 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-mnt2m_b35625a1-4b85-4915-8ac0-97ff11979513/cert-manager-cainjector/0.log" Jan 22 12:09:18 crc kubenswrapper[4874]: I0122 12:09:18.477288 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-cvlzs_bef508e9-b648-4c62-bc0d-91bf604067ed/cert-manager-webhook/0.log" Jan 22 12:09:24 crc kubenswrapper[4874]: I0122 12:09:24.339017 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-w7v89_71a74c6d-7ff3-4a3a-8c9d-bfa212b381b2/prometheus-operator/0.log" Jan 22 12:09:24 crc kubenswrapper[4874]: I0122 12:09:24.350922 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr_7a18bf90-2af7-4296-9416-1368e89a8a03/prometheus-operator-admission-webhook/0.log" Jan 22 12:09:24 crc kubenswrapper[4874]: I0122 12:09:24.365153 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8_60c00b52-d6b8-440e-9e60-966b44a87577/prometheus-operator-admission-webhook/0.log" Jan 22 12:09:24 crc kubenswrapper[4874]: I0122 12:09:24.389846 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-jblsg_edd7744e-1336-486a-90de-6568bae7f788/operator/0.log" Jan 22 12:09:24 crc kubenswrapper[4874]: I0122 12:09:24.407221 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-kqrkd_97207973-f6a4-4068-b777-c964c12092fd/perses-operator/0.log" Jan 22 12:09:24 crc kubenswrapper[4874]: I0122 12:09:24.715873 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:09:24 crc kubenswrapper[4874]: E0122 12:09:24.716092 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:09:29 crc kubenswrapper[4874]: I0122 12:09:29.892352 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5_9a48ea4e-8d83-4484-8cab-e9a38e86a2e1/extract/0.log" Jan 22 12:09:29 crc kubenswrapper[4874]: I0122 12:09:29.910213 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5_9a48ea4e-8d83-4484-8cab-e9a38e86a2e1/util/0.log" Jan 22 12:09:29 crc kubenswrapper[4874]: I0122 12:09:29.948128 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931az8rt5_9a48ea4e-8d83-4484-8cab-e9a38e86a2e1/pull/0.log" Jan 22 12:09:29 crc kubenswrapper[4874]: I0122 12:09:29.958374 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf_4db7503f-969b-4507-81bc-fb6ae0579495/extract/0.log" Jan 22 12:09:29 crc kubenswrapper[4874]: I0122 12:09:29.968780 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf_4db7503f-969b-4507-81bc-fb6ae0579495/util/0.log" Jan 22 12:09:29 crc kubenswrapper[4874]: I0122 12:09:29.977774 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn86tf_4db7503f-969b-4507-81bc-fb6ae0579495/pull/0.log" Jan 22 12:09:29 crc kubenswrapper[4874]: I0122 12:09:29.989746 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7_adad4abb-e416-4993-a166-0ac45ac75521/extract/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.000985 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7_adad4abb-e416-4993-a166-0ac45ac75521/util/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.009212 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ephtv7_adad4abb-e416-4993-a166-0ac45ac75521/pull/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.024844 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk_9634e8c9-9571-45b6-ad7f-c7d68d40c75a/extract/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.038411 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk_9634e8c9-9571-45b6-ad7f-c7d68d40c75a/util/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.045387 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f087fnfk_9634e8c9-9571-45b6-ad7f-c7d68d40c75a/pull/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.194678 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-98926_4c478499-7674-4b38-b0ea-15b5d9bc4702/registry-server/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.199580 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-98926_4c478499-7674-4b38-b0ea-15b5d9bc4702/extract-utilities/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.206686 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-98926_4c478499-7674-4b38-b0ea-15b5d9bc4702/extract-content/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.550937 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-g952h_1b32604b-2210-43c4-9b45-a833b8bdff64/registry-server/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.555434 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-g952h_1b32604b-2210-43c4-9b45-a833b8bdff64/extract-utilities/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.561033 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-g952h_1b32604b-2210-43c4-9b45-a833b8bdff64/extract-content/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.573215 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-st8lg_bb25a3d3-60b1-43ae-b007-19b20c362414/marketplace-operator/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.842424 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7qhz5_0684b3da-a3b2-496b-be2e-b8e0fe8e1277/registry-server/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.847677 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7qhz5_0684b3da-a3b2-496b-be2e-b8e0fe8e1277/extract-utilities/0.log" Jan 22 12:09:30 crc kubenswrapper[4874]: I0122 12:09:30.854875 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7qhz5_0684b3da-a3b2-496b-be2e-b8e0fe8e1277/extract-content/0.log" Jan 22 12:09:35 crc kubenswrapper[4874]: I0122 12:09:35.413106 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-w7v89_71a74c6d-7ff3-4a3a-8c9d-bfa212b381b2/prometheus-operator/0.log" Jan 22 12:09:35 crc kubenswrapper[4874]: I0122 12:09:35.424864 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr_7a18bf90-2af7-4296-9416-1368e89a8a03/prometheus-operator-admission-webhook/0.log" Jan 22 12:09:35 crc kubenswrapper[4874]: I0122 12:09:35.436773 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8_60c00b52-d6b8-440e-9e60-966b44a87577/prometheus-operator-admission-webhook/0.log" Jan 22 12:09:35 crc kubenswrapper[4874]: I0122 12:09:35.453243 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-jblsg_edd7744e-1336-486a-90de-6568bae7f788/operator/0.log" Jan 22 12:09:35 crc kubenswrapper[4874]: I0122 12:09:35.469122 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-kqrkd_97207973-f6a4-4068-b777-c964c12092fd/perses-operator/0.log" Jan 22 12:09:39 crc kubenswrapper[4874]: I0122 12:09:39.716723 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:09:39 crc kubenswrapper[4874]: E0122 12:09:39.717519 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:09:45 crc kubenswrapper[4874]: I0122 12:09:45.364127 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-w7v89_71a74c6d-7ff3-4a3a-8c9d-bfa212b381b2/prometheus-operator/0.log" Jan 22 12:09:45 crc kubenswrapper[4874]: I0122 12:09:45.398861 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6d95b449b4-2rphr_7a18bf90-2af7-4296-9416-1368e89a8a03/prometheus-operator-admission-webhook/0.log" Jan 22 12:09:45 crc kubenswrapper[4874]: I0122 12:09:45.409241 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6d95b449b4-d8hs8_60c00b52-d6b8-440e-9e60-966b44a87577/prometheus-operator-admission-webhook/0.log" Jan 22 12:09:45 crc kubenswrapper[4874]: I0122 12:09:45.427884 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-jblsg_edd7744e-1336-486a-90de-6568bae7f788/operator/0.log" Jan 22 12:09:45 crc kubenswrapper[4874]: I0122 12:09:45.437074 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-kqrkd_97207973-f6a4-4068-b777-c964c12092fd/perses-operator/0.log" Jan 22 12:09:45 crc kubenswrapper[4874]: I0122 12:09:45.533627 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-7nhkf_50438a7c-3a86-4aeb-a9c6-10ccb68f4593/cert-manager-controller/0.log" Jan 22 12:09:45 crc kubenswrapper[4874]: I0122 12:09:45.550916 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-mnt2m_b35625a1-4b85-4915-8ac0-97ff11979513/cert-manager-cainjector/0.log" Jan 22 12:09:45 crc kubenswrapper[4874]: I0122 12:09:45.561406 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-cvlzs_bef508e9-b648-4c62-bc0d-91bf604067ed/cert-manager-webhook/0.log" Jan 22 12:09:46 crc kubenswrapper[4874]: I0122 12:09:46.024600 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-7nhkf_50438a7c-3a86-4aeb-a9c6-10ccb68f4593/cert-manager-controller/0.log" Jan 22 12:09:46 crc kubenswrapper[4874]: I0122 12:09:46.047382 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-mnt2m_b35625a1-4b85-4915-8ac0-97ff11979513/cert-manager-cainjector/0.log" Jan 22 12:09:46 crc kubenswrapper[4874]: I0122 12:09:46.055482 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-cvlzs_bef508e9-b648-4c62-bc0d-91bf604067ed/cert-manager-webhook/0.log" Jan 22 12:09:46 crc kubenswrapper[4874]: I0122 12:09:46.582787 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6jqjk_a785e748-a557-4c24-8a4d-dc03cfc3c357/control-plane-machine-set-operator/0.log" Jan 22 12:09:46 crc kubenswrapper[4874]: I0122 12:09:46.597641 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wmwcz_9bec261f-fd1e-44f7-a402-cae34f722b6c/kube-rbac-proxy/0.log" Jan 22 12:09:46 crc kubenswrapper[4874]: I0122 12:09:46.605528 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wmwcz_9bec261f-fd1e-44f7-a402-cae34f722b6c/machine-api-operator/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.214474 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_30722d93-5804-4aba-a6e3-e2b891356163/alertmanager/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.221880 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_30722d93-5804-4aba-a6e3-e2b891356163/config-reloader/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.233215 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_30722d93-5804-4aba-a6e3-e2b891356163/oauth-proxy/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.240855 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_30722d93-5804-4aba-a6e3-e2b891356163/init-config-reloader/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.252137 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_curl_cd23c25e-9a15-4e31-be24-5d227c6ed353/curl/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.264387 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9_01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3/bridge/2.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.264543 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9_01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3/bridge/1.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.270258 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-6d59d88f55-dg4t9_01e2e1fb-0e3b-4c3b-b5c5-df459233c6e3/sg-core/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.284607 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np_1cca42d2-5390-4dc0-90fd-6abef972f828/oauth-proxy/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.290532 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np_1cca42d2-5390-4dc0-90fd-6abef972f828/bridge/2.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.290810 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np_1cca42d2-5390-4dc0-90fd-6abef972f828/bridge/1.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.299963 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-hv6np_1cca42d2-5390-4dc0-90fd-6abef972f828/sg-core/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.314262 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7_79746e6a-bee0-42eb-840c-36075d04abde/bridge/1.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.314837 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7_79746e6a-bee0-42eb-840c-36075d04abde/bridge/2.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.320868 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-684996bbcd-t6hb7_79746e6a-bee0-42eb-840c-36075d04abde/sg-core/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.331367 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg_16eb2f20-a170-4504-8601-a08d7174edef/oauth-proxy/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.338272 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg_16eb2f20-a170-4504-8601-a08d7174edef/bridge/2.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.338494 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg_16eb2f20-a170-4504-8601-a08d7174edef/bridge/1.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.343893 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-h9sbg_16eb2f20-a170-4504-8601-a08d7174edef/sg-core/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.356039 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf_7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2/oauth-proxy/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.361466 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf_7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2/bridge/1.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.361511 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf_7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2/bridge/2.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.367031 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-cx5rf_7c0b7de5-b0a4-4c4d-93c3-1feba19a95a2/sg-core/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.388207 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-interconnect-68864d46cb-jcvt5_da6dd65e-8b70-41a1-8a38-83392c65aef3/default-interconnect/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.395840 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-9qcrx_a14f7a5d-1819-4ca2-8683-96a338c70df6/prometheus-webhook-snmp/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.428154 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elastic-operator-864f7dd768-szq44_b54bc0d4-cf91-46af-bc4b-ea963cbd59bf/manager/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.459357 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_c9fa7f48-985a-46b3-a7dc-1b39dfc14243/elasticsearch/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.467529 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_c9fa7f48-985a-46b3-a7dc-1b39dfc14243/elastic-internal-init-filesystem/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.478020 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_c9fa7f48-985a-46b3-a7dc-1b39dfc14243/elastic-internal-suspend/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.497284 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_interconnect-operator-5bb49f789d-p9brk_cd733750-1915-4813-9e44-ec3777ce9c53/interconnect-operator/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.513228 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_7cd2b3c5-32a6-49c1-984c-86aa8fe36f85/prometheus/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.521781 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_7cd2b3c5-32a6-49c1-984c-86aa8fe36f85/config-reloader/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.533584 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_7cd2b3c5-32a6-49c1-984c-86aa8fe36f85/oauth-proxy/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.539627 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_7cd2b3c5-32a6-49c1-984c-86aa8fe36f85/init-config-reloader/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.574849 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_0930f68b-3161-43da-80c9-440cf31d98b9/docker-build/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.579467 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_0930f68b-3161-43da-80c9-440cf31d98b9/git-clone/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.586764 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_0930f68b-3161-43da-80c9-440cf31d98b9/manage-dockerfile/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.604788 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_qdr-test_056ff179-9e03-4f48-b3d3-fa995d0c4093/qdr/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.645072 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_b6347418-07a3-41af-aea9-1eddb77e64fb/docker-build/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.703054 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_b6347418-07a3-41af-aea9-1eddb77e64fb/git-clone/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.711478 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_b6347418-07a3-41af-aea9-1eddb77e64fb/manage-dockerfile/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.879980 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-66768bfd46-6v9td_9c1cf40e-37fa-4e5a-8bb3-8ee8aee9c413/operator/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.921744 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-2-build_c087bdb2-7a13-44f4-a3e8-d023752790b1/docker-build/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.926839 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-2-build_c087bdb2-7a13-44f4-a3e8-d023752790b1/git-clone/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.939220 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-2-build_c087bdb2-7a13-44f4-a3e8-d023752790b1/manage-dockerfile/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.975360 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-2-build_d9c696c3-8c17-45a3-93d2-75801fa0bff4/docker-build/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.981046 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-2-build_d9c696c3-8c17-45a3-93d2-75801fa0bff4/git-clone/0.log" Jan 22 12:09:47 crc kubenswrapper[4874]: I0122 12:09:47.987818 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-2-build_d9c696c3-8c17-45a3-93d2-75801fa0bff4/manage-dockerfile/0.log" Jan 22 12:09:48 crc kubenswrapper[4874]: I0122 12:09:48.054551 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-2-build_e36b7db7-fd44-4552-8710-adb858c931c9/docker-build/0.log" Jan 22 12:09:48 crc kubenswrapper[4874]: I0122 12:09:48.059570 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-2-build_e36b7db7-fd44-4552-8710-adb858c931c9/git-clone/0.log" Jan 22 12:09:48 crc kubenswrapper[4874]: I0122 12:09:48.067094 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-2-build_e36b7db7-fd44-4552-8710-adb858c931c9/manage-dockerfile/0.log" Jan 22 12:09:50 crc kubenswrapper[4874]: I0122 12:09:50.716299 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:09:50 crc kubenswrapper[4874]: E0122 12:09:50.716819 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:09:51 crc kubenswrapper[4874]: I0122 12:09:51.149650 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-779b7b47dc-4zkhw_f01ecec4-43d6-424e-843b-ef89a732199e/operator/0.log" Jan 22 12:09:51 crc kubenswrapper[4874]: I0122 12:09:51.166022 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-fpm2k_91078bec-5716-4d9f-ade0-53de332b1871/smoketest-collectd/0.log" Jan 22 12:09:51 crc kubenswrapper[4874]: I0122 12:09:51.171179 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-fpm2k_91078bec-5716-4d9f-ade0-53de332b1871/smoketest-ceilometer/0.log" Jan 22 12:09:51 crc kubenswrapper[4874]: I0122 12:09:51.186859 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-qb5kr_c1f3f1ed-6e23-4ad5-877c-522b9cb42ade/smoketest-collectd/0.log" Jan 22 12:09:51 crc kubenswrapper[4874]: I0122 12:09:51.191307 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-qb5kr_c1f3f1ed-6e23-4ad5-877c-522b9cb42ade/smoketest-ceilometer/0.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.569114 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pdb2m_0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4/kube-multus-additional-cni-plugins/0.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.577662 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pdb2m_0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4/egress-router-binary-copy/0.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.586807 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pdb2m_0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4/cni-plugins/0.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.594969 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pdb2m_0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4/bond-cni-plugin/0.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.603587 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pdb2m_0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4/routeoverride-cni/0.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.611787 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pdb2m_0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4/whereabouts-cni-bincopy/0.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.623696 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pdb2m_0dbaa8a2-585e-4ab7-a8f5-7b224ad8c5c4/whereabouts-cni/0.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.636675 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-f64ps_1795c220-db74-434e-9111-917ff6d95077/multus-admission-controller/0.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.644567 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-f64ps_1795c220-db74-434e-9111-917ff6d95077/kube-rbac-proxy/0.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.670862 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/3.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.685960 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-krrtc_977746b5-ac1b-4b6e-bdbc-ddd90225e68c/kube-multus/2.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.705577 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-lr2vd_5451fbab-ebad-42e7-bb80-f94bad10d571/network-metrics-daemon/0.log" Jan 22 12:09:52 crc kubenswrapper[4874]: I0122 12:09:52.710769 4874 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-lr2vd_5451fbab-ebad-42e7-bb80-f94bad10d571/kube-rbac-proxy/0.log" Jan 22 12:10:05 crc kubenswrapper[4874]: I0122 12:10:05.715767 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:10:05 crc kubenswrapper[4874]: E0122 12:10:05.716623 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:10:19 crc kubenswrapper[4874]: I0122 12:10:19.717310 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:10:19 crc kubenswrapper[4874]: E0122 12:10:19.718383 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:10:30 crc kubenswrapper[4874]: I0122 12:10:30.716727 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:10:30 crc kubenswrapper[4874]: E0122 12:10:30.717800 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:10:42 crc kubenswrapper[4874]: I0122 12:10:42.716929 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:10:42 crc kubenswrapper[4874]: E0122 12:10:42.717804 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:10:54 crc kubenswrapper[4874]: I0122 12:10:54.717155 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:10:54 crc kubenswrapper[4874]: E0122 12:10:54.717872 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:11:08 crc kubenswrapper[4874]: I0122 12:11:08.719031 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:11:08 crc kubenswrapper[4874]: E0122 12:11:08.719794 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:11:21 crc kubenswrapper[4874]: I0122 12:11:21.715999 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:11:21 crc kubenswrapper[4874]: E0122 12:11:21.716774 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:11:33 crc kubenswrapper[4874]: I0122 12:11:33.716913 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:11:33 crc kubenswrapper[4874]: E0122 12:11:33.718045 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:11:46 crc kubenswrapper[4874]: I0122 12:11:46.728341 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:11:46 crc kubenswrapper[4874]: I0122 12:11:46.992090 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"4bd591f21c90200802db0c3a40e96e2477cc7ef98ac950449bf351a75f78ca92"} Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.751701 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h5v4d"] Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.754798 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.761874 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h5v4d"] Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.810180 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-utilities\") pod \"certified-operators-h5v4d\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.810311 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-catalog-content\") pod \"certified-operators-h5v4d\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.810441 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mbxb\" (UniqueName: \"kubernetes.io/projected/9bf24c61-16db-4593-ba56-3e13cf4b2320-kube-api-access-2mbxb\") pod \"certified-operators-h5v4d\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.912137 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-utilities\") pod \"certified-operators-h5v4d\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.912474 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-catalog-content\") pod \"certified-operators-h5v4d\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.912647 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mbxb\" (UniqueName: \"kubernetes.io/projected/9bf24c61-16db-4593-ba56-3e13cf4b2320-kube-api-access-2mbxb\") pod \"certified-operators-h5v4d\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.913692 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-utilities\") pod \"certified-operators-h5v4d\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.914132 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-catalog-content\") pod \"certified-operators-h5v4d\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:35 crc kubenswrapper[4874]: I0122 12:12:35.952004 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mbxb\" (UniqueName: \"kubernetes.io/projected/9bf24c61-16db-4593-ba56-3e13cf4b2320-kube-api-access-2mbxb\") pod \"certified-operators-h5v4d\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:36 crc kubenswrapper[4874]: I0122 12:12:36.087822 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:36 crc kubenswrapper[4874]: I0122 12:12:36.344075 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h5v4d"] Jan 22 12:12:36 crc kubenswrapper[4874]: I0122 12:12:36.479503 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5v4d" event={"ID":"9bf24c61-16db-4593-ba56-3e13cf4b2320","Type":"ContainerStarted","Data":"a35d385e2fc22970f70e33cd6e3ffc93c10baa2c9a892e4f5c5d067046cfdb4f"} Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.490442 4874 generic.go:334] "Generic (PLEG): container finished" podID="9bf24c61-16db-4593-ba56-3e13cf4b2320" containerID="9d0709b48d1811b8c84b55687b33409b330c566b7d4ee4ab5f7927d6bece36d9" exitCode=0 Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.490498 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5v4d" event={"ID":"9bf24c61-16db-4593-ba56-3e13cf4b2320","Type":"ContainerDied","Data":"9d0709b48d1811b8c84b55687b33409b330c566b7d4ee4ab5f7927d6bece36d9"} Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.492146 4874 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.575088 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9ftt2"] Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.577540 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.588702 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9ftt2"] Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.644253 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-catalog-content\") pod \"redhat-operators-9ftt2\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.644321 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f54p\" (UniqueName: \"kubernetes.io/projected/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-kube-api-access-5f54p\") pod \"redhat-operators-9ftt2\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.644361 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-utilities\") pod \"redhat-operators-9ftt2\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.745419 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-catalog-content\") pod \"redhat-operators-9ftt2\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.745499 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f54p\" (UniqueName: \"kubernetes.io/projected/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-kube-api-access-5f54p\") pod \"redhat-operators-9ftt2\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.745540 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-utilities\") pod \"redhat-operators-9ftt2\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.746195 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-catalog-content\") pod \"redhat-operators-9ftt2\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.746323 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-utilities\") pod \"redhat-operators-9ftt2\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.771973 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f54p\" (UniqueName: \"kubernetes.io/projected/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-kube-api-access-5f54p\") pod \"redhat-operators-9ftt2\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:37 crc kubenswrapper[4874]: I0122 12:12:37.901941 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:38 crc kubenswrapper[4874]: I0122 12:12:38.339106 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9ftt2"] Jan 22 12:12:38 crc kubenswrapper[4874]: W0122 12:12:38.348798 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64ecc5d7_a8b0_4e2b_9520_f74ef155f54a.slice/crio-305cd87b6ec882e88fb40a1ac4fd939e5ee3a05f35521146bfa61373338404dc WatchSource:0}: Error finding container 305cd87b6ec882e88fb40a1ac4fd939e5ee3a05f35521146bfa61373338404dc: Status 404 returned error can't find the container with id 305cd87b6ec882e88fb40a1ac4fd939e5ee3a05f35521146bfa61373338404dc Jan 22 12:12:38 crc kubenswrapper[4874]: I0122 12:12:38.499238 4874 generic.go:334] "Generic (PLEG): container finished" podID="9bf24c61-16db-4593-ba56-3e13cf4b2320" containerID="f6de90acdaad3f5455174ad4b6c2480f83ce787bdf0eeb279fa155ce857ef467" exitCode=0 Jan 22 12:12:38 crc kubenswrapper[4874]: I0122 12:12:38.499312 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5v4d" event={"ID":"9bf24c61-16db-4593-ba56-3e13cf4b2320","Type":"ContainerDied","Data":"f6de90acdaad3f5455174ad4b6c2480f83ce787bdf0eeb279fa155ce857ef467"} Jan 22 12:12:38 crc kubenswrapper[4874]: I0122 12:12:38.501105 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftt2" event={"ID":"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a","Type":"ContainerStarted","Data":"ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496"} Jan 22 12:12:38 crc kubenswrapper[4874]: I0122 12:12:38.501142 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftt2" event={"ID":"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a","Type":"ContainerStarted","Data":"305cd87b6ec882e88fb40a1ac4fd939e5ee3a05f35521146bfa61373338404dc"} Jan 22 12:12:39 crc kubenswrapper[4874]: I0122 12:12:39.521127 4874 generic.go:334] "Generic (PLEG): container finished" podID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerID="ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496" exitCode=0 Jan 22 12:12:39 crc kubenswrapper[4874]: I0122 12:12:39.521160 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftt2" event={"ID":"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a","Type":"ContainerDied","Data":"ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496"} Jan 22 12:12:39 crc kubenswrapper[4874]: I0122 12:12:39.521549 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftt2" event={"ID":"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a","Type":"ContainerStarted","Data":"c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285"} Jan 22 12:12:39 crc kubenswrapper[4874]: I0122 12:12:39.533568 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5v4d" event={"ID":"9bf24c61-16db-4593-ba56-3e13cf4b2320","Type":"ContainerStarted","Data":"ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442"} Jan 22 12:12:39 crc kubenswrapper[4874]: I0122 12:12:39.565249 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h5v4d" podStartSLOduration=3.192098301 podStartE2EDuration="4.565225182s" podCreationTimestamp="2026-01-22 12:12:35 +0000 UTC" firstStartedPulling="2026-01-22 12:12:37.491900233 +0000 UTC m=+1931.336971313" lastFinishedPulling="2026-01-22 12:12:38.865027114 +0000 UTC m=+1932.710098194" observedRunningTime="2026-01-22 12:12:39.561257119 +0000 UTC m=+1933.406328209" watchObservedRunningTime="2026-01-22 12:12:39.565225182 +0000 UTC m=+1933.410296292" Jan 22 12:12:43 crc kubenswrapper[4874]: I0122 12:12:43.574695 4874 generic.go:334] "Generic (PLEG): container finished" podID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerID="c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285" exitCode=0 Jan 22 12:12:43 crc kubenswrapper[4874]: I0122 12:12:43.574738 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftt2" event={"ID":"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a","Type":"ContainerDied","Data":"c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285"} Jan 22 12:12:44 crc kubenswrapper[4874]: I0122 12:12:44.586571 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftt2" event={"ID":"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a","Type":"ContainerStarted","Data":"c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a"} Jan 22 12:12:44 crc kubenswrapper[4874]: I0122 12:12:44.618523 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9ftt2" podStartSLOduration=1.781040654 podStartE2EDuration="7.618505431s" podCreationTimestamp="2026-01-22 12:12:37 +0000 UTC" firstStartedPulling="2026-01-22 12:12:38.504635547 +0000 UTC m=+1932.349706617" lastFinishedPulling="2026-01-22 12:12:44.342100324 +0000 UTC m=+1938.187171394" observedRunningTime="2026-01-22 12:12:44.614556938 +0000 UTC m=+1938.459628058" watchObservedRunningTime="2026-01-22 12:12:44.618505431 +0000 UTC m=+1938.463576511" Jan 22 12:12:46 crc kubenswrapper[4874]: I0122 12:12:46.088265 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:46 crc kubenswrapper[4874]: I0122 12:12:46.088361 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:46 crc kubenswrapper[4874]: I0122 12:12:46.164356 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:46 crc kubenswrapper[4874]: I0122 12:12:46.673258 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:46 crc kubenswrapper[4874]: I0122 12:12:46.927908 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h5v4d"] Jan 22 12:12:47 crc kubenswrapper[4874]: I0122 12:12:47.902849 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:47 crc kubenswrapper[4874]: I0122 12:12:47.902923 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:48 crc kubenswrapper[4874]: I0122 12:12:48.617956 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h5v4d" podUID="9bf24c61-16db-4593-ba56-3e13cf4b2320" containerName="registry-server" containerID="cri-o://ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442" gracePeriod=2 Jan 22 12:12:48 crc kubenswrapper[4874]: I0122 12:12:48.961318 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9ftt2" podUID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerName="registry-server" probeResult="failure" output=< Jan 22 12:12:48 crc kubenswrapper[4874]: timeout: failed to connect service ":50051" within 1s Jan 22 12:12:48 crc kubenswrapper[4874]: > Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.050646 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.129507 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mbxb\" (UniqueName: \"kubernetes.io/projected/9bf24c61-16db-4593-ba56-3e13cf4b2320-kube-api-access-2mbxb\") pod \"9bf24c61-16db-4593-ba56-3e13cf4b2320\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.129570 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-utilities\") pod \"9bf24c61-16db-4593-ba56-3e13cf4b2320\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.129670 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-catalog-content\") pod \"9bf24c61-16db-4593-ba56-3e13cf4b2320\" (UID: \"9bf24c61-16db-4593-ba56-3e13cf4b2320\") " Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.130913 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-utilities" (OuterVolumeSpecName: "utilities") pod "9bf24c61-16db-4593-ba56-3e13cf4b2320" (UID: "9bf24c61-16db-4593-ba56-3e13cf4b2320"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.150889 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bf24c61-16db-4593-ba56-3e13cf4b2320-kube-api-access-2mbxb" (OuterVolumeSpecName: "kube-api-access-2mbxb") pod "9bf24c61-16db-4593-ba56-3e13cf4b2320" (UID: "9bf24c61-16db-4593-ba56-3e13cf4b2320"). InnerVolumeSpecName "kube-api-access-2mbxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.180156 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9bf24c61-16db-4593-ba56-3e13cf4b2320" (UID: "9bf24c61-16db-4593-ba56-3e13cf4b2320"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.231462 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mbxb\" (UniqueName: \"kubernetes.io/projected/9bf24c61-16db-4593-ba56-3e13cf4b2320-kube-api-access-2mbxb\") on node \"crc\" DevicePath \"\"" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.231505 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.231517 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf24c61-16db-4593-ba56-3e13cf4b2320-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.626111 4874 generic.go:334] "Generic (PLEG): container finished" podID="9bf24c61-16db-4593-ba56-3e13cf4b2320" containerID="ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442" exitCode=0 Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.626156 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5v4d" event={"ID":"9bf24c61-16db-4593-ba56-3e13cf4b2320","Type":"ContainerDied","Data":"ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442"} Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.626186 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5v4d" event={"ID":"9bf24c61-16db-4593-ba56-3e13cf4b2320","Type":"ContainerDied","Data":"a35d385e2fc22970f70e33cd6e3ffc93c10baa2c9a892e4f5c5d067046cfdb4f"} Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.626206 4874 scope.go:117] "RemoveContainer" containerID="ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.626362 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5v4d" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.654432 4874 scope.go:117] "RemoveContainer" containerID="f6de90acdaad3f5455174ad4b6c2480f83ce787bdf0eeb279fa155ce857ef467" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.665575 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h5v4d"] Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.670646 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h5v4d"] Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.683497 4874 scope.go:117] "RemoveContainer" containerID="9d0709b48d1811b8c84b55687b33409b330c566b7d4ee4ab5f7927d6bece36d9" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.709064 4874 scope.go:117] "RemoveContainer" containerID="ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442" Jan 22 12:12:49 crc kubenswrapper[4874]: E0122 12:12:49.709545 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442\": container with ID starting with ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442 not found: ID does not exist" containerID="ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.709610 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442"} err="failed to get container status \"ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442\": rpc error: code = NotFound desc = could not find container \"ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442\": container with ID starting with ac751bd5d37eaf7b3f40e2da3e07f8429036c76a792a36f2be475b0a48fa4442 not found: ID does not exist" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.711142 4874 scope.go:117] "RemoveContainer" containerID="f6de90acdaad3f5455174ad4b6c2480f83ce787bdf0eeb279fa155ce857ef467" Jan 22 12:12:49 crc kubenswrapper[4874]: E0122 12:12:49.713553 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6de90acdaad3f5455174ad4b6c2480f83ce787bdf0eeb279fa155ce857ef467\": container with ID starting with f6de90acdaad3f5455174ad4b6c2480f83ce787bdf0eeb279fa155ce857ef467 not found: ID does not exist" containerID="f6de90acdaad3f5455174ad4b6c2480f83ce787bdf0eeb279fa155ce857ef467" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.713631 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6de90acdaad3f5455174ad4b6c2480f83ce787bdf0eeb279fa155ce857ef467"} err="failed to get container status \"f6de90acdaad3f5455174ad4b6c2480f83ce787bdf0eeb279fa155ce857ef467\": rpc error: code = NotFound desc = could not find container \"f6de90acdaad3f5455174ad4b6c2480f83ce787bdf0eeb279fa155ce857ef467\": container with ID starting with f6de90acdaad3f5455174ad4b6c2480f83ce787bdf0eeb279fa155ce857ef467 not found: ID does not exist" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.713660 4874 scope.go:117] "RemoveContainer" containerID="9d0709b48d1811b8c84b55687b33409b330c566b7d4ee4ab5f7927d6bece36d9" Jan 22 12:12:49 crc kubenswrapper[4874]: E0122 12:12:49.714075 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d0709b48d1811b8c84b55687b33409b330c566b7d4ee4ab5f7927d6bece36d9\": container with ID starting with 9d0709b48d1811b8c84b55687b33409b330c566b7d4ee4ab5f7927d6bece36d9 not found: ID does not exist" containerID="9d0709b48d1811b8c84b55687b33409b330c566b7d4ee4ab5f7927d6bece36d9" Jan 22 12:12:49 crc kubenswrapper[4874]: I0122 12:12:49.714160 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d0709b48d1811b8c84b55687b33409b330c566b7d4ee4ab5f7927d6bece36d9"} err="failed to get container status \"9d0709b48d1811b8c84b55687b33409b330c566b7d4ee4ab5f7927d6bece36d9\": rpc error: code = NotFound desc = could not find container \"9d0709b48d1811b8c84b55687b33409b330c566b7d4ee4ab5f7927d6bece36d9\": container with ID starting with 9d0709b48d1811b8c84b55687b33409b330c566b7d4ee4ab5f7927d6bece36d9 not found: ID does not exist" Jan 22 12:12:50 crc kubenswrapper[4874]: I0122 12:12:50.728952 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bf24c61-16db-4593-ba56-3e13cf4b2320" path="/var/lib/kubelet/pods/9bf24c61-16db-4593-ba56-3e13cf4b2320/volumes" Jan 22 12:12:57 crc kubenswrapper[4874]: I0122 12:12:57.983184 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:58 crc kubenswrapper[4874]: I0122 12:12:58.053051 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:12:58 crc kubenswrapper[4874]: I0122 12:12:58.443887 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9ftt2"] Jan 22 12:12:59 crc kubenswrapper[4874]: I0122 12:12:59.724735 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9ftt2" podUID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerName="registry-server" containerID="cri-o://c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a" gracePeriod=2 Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.164355 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.230186 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-utilities\") pod \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.230288 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5f54p\" (UniqueName: \"kubernetes.io/projected/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-kube-api-access-5f54p\") pod \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.230317 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-catalog-content\") pod \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\" (UID: \"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a\") " Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.231273 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-utilities" (OuterVolumeSpecName: "utilities") pod "64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" (UID: "64ecc5d7-a8b0-4e2b-9520-f74ef155f54a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.250415 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-kube-api-access-5f54p" (OuterVolumeSpecName: "kube-api-access-5f54p") pod "64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" (UID: "64ecc5d7-a8b0-4e2b-9520-f74ef155f54a"). InnerVolumeSpecName "kube-api-access-5f54p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.331509 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.331549 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5f54p\" (UniqueName: \"kubernetes.io/projected/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-kube-api-access-5f54p\") on node \"crc\" DevicePath \"\"" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.348734 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" (UID: "64ecc5d7-a8b0-4e2b-9520-f74ef155f54a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.432463 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.758429 4874 generic.go:334] "Generic (PLEG): container finished" podID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerID="c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a" exitCode=0 Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.758756 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftt2" event={"ID":"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a","Type":"ContainerDied","Data":"c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a"} Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.758799 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftt2" event={"ID":"64ecc5d7-a8b0-4e2b-9520-f74ef155f54a","Type":"ContainerDied","Data":"305cd87b6ec882e88fb40a1ac4fd939e5ee3a05f35521146bfa61373338404dc"} Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.758827 4874 scope.go:117] "RemoveContainer" containerID="c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.759066 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ftt2" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.801859 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9ftt2"] Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.803444 4874 scope.go:117] "RemoveContainer" containerID="c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.811173 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9ftt2"] Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.833131 4874 scope.go:117] "RemoveContainer" containerID="ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.860312 4874 scope.go:117] "RemoveContainer" containerID="c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a" Jan 22 12:13:00 crc kubenswrapper[4874]: E0122 12:13:00.861170 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a\": container with ID starting with c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a not found: ID does not exist" containerID="c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.861213 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a"} err="failed to get container status \"c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a\": rpc error: code = NotFound desc = could not find container \"c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a\": container with ID starting with c86e8e87b6b18703131d3c91af4e04cccee578cc740be51fa2f7e27d62b9954a not found: ID does not exist" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.861239 4874 scope.go:117] "RemoveContainer" containerID="c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285" Jan 22 12:13:00 crc kubenswrapper[4874]: E0122 12:13:00.861877 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285\": container with ID starting with c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285 not found: ID does not exist" containerID="c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.861979 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285"} err="failed to get container status \"c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285\": rpc error: code = NotFound desc = could not find container \"c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285\": container with ID starting with c741cf9cde1c4861c16f849da921132c17a3beb19a123d25322d71547cf13285 not found: ID does not exist" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.862129 4874 scope.go:117] "RemoveContainer" containerID="ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496" Jan 22 12:13:00 crc kubenswrapper[4874]: E0122 12:13:00.862441 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496\": container with ID starting with ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496 not found: ID does not exist" containerID="ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496" Jan 22 12:13:00 crc kubenswrapper[4874]: I0122 12:13:00.862468 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496"} err="failed to get container status \"ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496\": rpc error: code = NotFound desc = could not find container \"ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496\": container with ID starting with ad5a3be4a4bb4d5d1a756d0a99a798f0e7b745ab428423fc92215233ed167496 not found: ID does not exist" Jan 22 12:13:02 crc kubenswrapper[4874]: I0122 12:13:02.732652 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" path="/var/lib/kubelet/pods/64ecc5d7-a8b0-4e2b-9520-f74ef155f54a/volumes" Jan 22 12:14:13 crc kubenswrapper[4874]: I0122 12:14:13.520286 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:14:13 crc kubenswrapper[4874]: I0122 12:14:13.522570 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:14:43 crc kubenswrapper[4874]: I0122 12:14:43.520743 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:14:43 crc kubenswrapper[4874]: I0122 12:14:43.521472 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.156023 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt"] Jan 22 12:15:00 crc kubenswrapper[4874]: E0122 12:15:00.157044 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bf24c61-16db-4593-ba56-3e13cf4b2320" containerName="extract-utilities" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.157069 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bf24c61-16db-4593-ba56-3e13cf4b2320" containerName="extract-utilities" Jan 22 12:15:00 crc kubenswrapper[4874]: E0122 12:15:00.157091 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerName="registry-server" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.157106 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerName="registry-server" Jan 22 12:15:00 crc kubenswrapper[4874]: E0122 12:15:00.157147 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bf24c61-16db-4593-ba56-3e13cf4b2320" containerName="extract-content" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.157161 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bf24c61-16db-4593-ba56-3e13cf4b2320" containerName="extract-content" Jan 22 12:15:00 crc kubenswrapper[4874]: E0122 12:15:00.157180 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerName="extract-content" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.157192 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerName="extract-content" Jan 22 12:15:00 crc kubenswrapper[4874]: E0122 12:15:00.157210 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bf24c61-16db-4593-ba56-3e13cf4b2320" containerName="registry-server" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.157223 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bf24c61-16db-4593-ba56-3e13cf4b2320" containerName="registry-server" Jan 22 12:15:00 crc kubenswrapper[4874]: E0122 12:15:00.157242 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerName="extract-utilities" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.157254 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerName="extract-utilities" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.157487 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bf24c61-16db-4593-ba56-3e13cf4b2320" containerName="registry-server" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.157525 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="64ecc5d7-a8b0-4e2b-9520-f74ef155f54a" containerName="registry-server" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.158269 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.163223 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.166237 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.179994 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt"] Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.214956 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-secret-volume\") pod \"collect-profiles-29484735-jrjzt\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.215008 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-config-volume\") pod \"collect-profiles-29484735-jrjzt\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.215079 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlmrd\" (UniqueName: \"kubernetes.io/projected/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-kube-api-access-tlmrd\") pod \"collect-profiles-29484735-jrjzt\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.316038 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-secret-volume\") pod \"collect-profiles-29484735-jrjzt\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.316087 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-config-volume\") pod \"collect-profiles-29484735-jrjzt\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.316126 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlmrd\" (UniqueName: \"kubernetes.io/projected/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-kube-api-access-tlmrd\") pod \"collect-profiles-29484735-jrjzt\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.317062 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-config-volume\") pod \"collect-profiles-29484735-jrjzt\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.337002 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-secret-volume\") pod \"collect-profiles-29484735-jrjzt\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.337006 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlmrd\" (UniqueName: \"kubernetes.io/projected/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-kube-api-access-tlmrd\") pod \"collect-profiles-29484735-jrjzt\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.477593 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:00 crc kubenswrapper[4874]: I0122 12:15:00.920461 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt"] Jan 22 12:15:01 crc kubenswrapper[4874]: I0122 12:15:01.930073 4874 generic.go:334] "Generic (PLEG): container finished" podID="1b771c8b-4e94-4d20-ac6b-72c4798a65cb" containerID="afe34a02bbd93802a374e833960c467dfba9da2664b1e4cf5e3b189ded4d6c1d" exitCode=0 Jan 22 12:15:01 crc kubenswrapper[4874]: I0122 12:15:01.930284 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" event={"ID":"1b771c8b-4e94-4d20-ac6b-72c4798a65cb","Type":"ContainerDied","Data":"afe34a02bbd93802a374e833960c467dfba9da2664b1e4cf5e3b189ded4d6c1d"} Jan 22 12:15:01 crc kubenswrapper[4874]: I0122 12:15:01.930309 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" event={"ID":"1b771c8b-4e94-4d20-ac6b-72c4798a65cb","Type":"ContainerStarted","Data":"e90b3e2345007d5f042662e62d34e7b1bedee54ad7ec531e5233bbfae279d163"} Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.222972 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.373973 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-config-volume\") pod \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.374071 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlmrd\" (UniqueName: \"kubernetes.io/projected/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-kube-api-access-tlmrd\") pod \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.374280 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-secret-volume\") pod \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\" (UID: \"1b771c8b-4e94-4d20-ac6b-72c4798a65cb\") " Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.375014 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "1b771c8b-4e94-4d20-ac6b-72c4798a65cb" (UID: "1b771c8b-4e94-4d20-ac6b-72c4798a65cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.376530 4874 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.380538 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1b771c8b-4e94-4d20-ac6b-72c4798a65cb" (UID: "1b771c8b-4e94-4d20-ac6b-72c4798a65cb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.381093 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-kube-api-access-tlmrd" (OuterVolumeSpecName: "kube-api-access-tlmrd") pod "1b771c8b-4e94-4d20-ac6b-72c4798a65cb" (UID: "1b771c8b-4e94-4d20-ac6b-72c4798a65cb"). InnerVolumeSpecName "kube-api-access-tlmrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.477694 4874 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.477755 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlmrd\" (UniqueName: \"kubernetes.io/projected/1b771c8b-4e94-4d20-ac6b-72c4798a65cb-kube-api-access-tlmrd\") on node \"crc\" DevicePath \"\"" Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.959390 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" event={"ID":"1b771c8b-4e94-4d20-ac6b-72c4798a65cb","Type":"ContainerDied","Data":"e90b3e2345007d5f042662e62d34e7b1bedee54ad7ec531e5233bbfae279d163"} Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.959450 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e90b3e2345007d5f042662e62d34e7b1bedee54ad7ec531e5233bbfae279d163" Jan 22 12:15:03 crc kubenswrapper[4874]: I0122 12:15:03.959470 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484735-jrjzt" Jan 22 12:15:04 crc kubenswrapper[4874]: I0122 12:15:04.294268 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx"] Jan 22 12:15:04 crc kubenswrapper[4874]: I0122 12:15:04.305885 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484690-npncx"] Jan 22 12:15:04 crc kubenswrapper[4874]: I0122 12:15:04.725079 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a15050a-7cbd-40f6-a656-a68293c0878a" path="/var/lib/kubelet/pods/0a15050a-7cbd-40f6-a656-a68293c0878a/volumes" Jan 22 12:15:13 crc kubenswrapper[4874]: I0122 12:15:13.520641 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:15:13 crc kubenswrapper[4874]: I0122 12:15:13.522424 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:15:13 crc kubenswrapper[4874]: I0122 12:15:13.522621 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 12:15:13 crc kubenswrapper[4874]: I0122 12:15:13.523459 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4bd591f21c90200802db0c3a40e96e2477cc7ef98ac950449bf351a75f78ca92"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 12:15:13 crc kubenswrapper[4874]: I0122 12:15:13.523634 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://4bd591f21c90200802db0c3a40e96e2477cc7ef98ac950449bf351a75f78ca92" gracePeriod=600 Jan 22 12:15:14 crc kubenswrapper[4874]: I0122 12:15:14.042356 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="4bd591f21c90200802db0c3a40e96e2477cc7ef98ac950449bf351a75f78ca92" exitCode=0 Jan 22 12:15:14 crc kubenswrapper[4874]: I0122 12:15:14.042425 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"4bd591f21c90200802db0c3a40e96e2477cc7ef98ac950449bf351a75f78ca92"} Jan 22 12:15:14 crc kubenswrapper[4874]: I0122 12:15:14.043120 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646"} Jan 22 12:15:14 crc kubenswrapper[4874]: I0122 12:15:14.043166 4874 scope.go:117] "RemoveContainer" containerID="9e9d8998fedf3ebcd297ebcad326b3f27048547f0407cb9bc7f2e574679a4b15" Jan 22 12:15:33 crc kubenswrapper[4874]: I0122 12:15:33.324230 4874 scope.go:117] "RemoveContainer" containerID="059406f38b7eefe033761eb54f07a9cd0f802e07f88ff68ccbdb582fae9d8d28" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.197847 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sgqzt"] Jan 22 12:16:19 crc kubenswrapper[4874]: E0122 12:16:19.198863 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b771c8b-4e94-4d20-ac6b-72c4798a65cb" containerName="collect-profiles" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.198885 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b771c8b-4e94-4d20-ac6b-72c4798a65cb" containerName="collect-profiles" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.199144 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b771c8b-4e94-4d20-ac6b-72c4798a65cb" containerName="collect-profiles" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.200686 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.208644 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sgqzt"] Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.289004 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-utilities\") pod \"community-operators-sgqzt\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.289170 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-catalog-content\") pod \"community-operators-sgqzt\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.289385 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzgsk\" (UniqueName: \"kubernetes.io/projected/d598bb98-3008-4081-8187-bbd8e622bb3a-kube-api-access-xzgsk\") pod \"community-operators-sgqzt\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.391260 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-catalog-content\") pod \"community-operators-sgqzt\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.391606 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzgsk\" (UniqueName: \"kubernetes.io/projected/d598bb98-3008-4081-8187-bbd8e622bb3a-kube-api-access-xzgsk\") pod \"community-operators-sgqzt\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.391674 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-utilities\") pod \"community-operators-sgqzt\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.392208 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-catalog-content\") pod \"community-operators-sgqzt\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.392608 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-utilities\") pod \"community-operators-sgqzt\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.412687 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzgsk\" (UniqueName: \"kubernetes.io/projected/d598bb98-3008-4081-8187-bbd8e622bb3a-kube-api-access-xzgsk\") pod \"community-operators-sgqzt\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:19 crc kubenswrapper[4874]: I0122 12:16:19.536120 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:20 crc kubenswrapper[4874]: I0122 12:16:20.084884 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sgqzt"] Jan 22 12:16:20 crc kubenswrapper[4874]: W0122 12:16:20.089539 4874 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd598bb98_3008_4081_8187_bbd8e622bb3a.slice/crio-095c3d0baf601af83872c20e63e2bb7f4dd45902aba137ead9517326a397037a WatchSource:0}: Error finding container 095c3d0baf601af83872c20e63e2bb7f4dd45902aba137ead9517326a397037a: Status 404 returned error can't find the container with id 095c3d0baf601af83872c20e63e2bb7f4dd45902aba137ead9517326a397037a Jan 22 12:16:20 crc kubenswrapper[4874]: I0122 12:16:20.609001 4874 generic.go:334] "Generic (PLEG): container finished" podID="d598bb98-3008-4081-8187-bbd8e622bb3a" containerID="31aa3e5fdae92637dce6ea824242e7b91959d9977ef5dd71d07c2a22cad8d5de" exitCode=0 Jan 22 12:16:20 crc kubenswrapper[4874]: I0122 12:16:20.609038 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgqzt" event={"ID":"d598bb98-3008-4081-8187-bbd8e622bb3a","Type":"ContainerDied","Data":"31aa3e5fdae92637dce6ea824242e7b91959d9977ef5dd71d07c2a22cad8d5de"} Jan 22 12:16:20 crc kubenswrapper[4874]: I0122 12:16:20.609063 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgqzt" event={"ID":"d598bb98-3008-4081-8187-bbd8e622bb3a","Type":"ContainerStarted","Data":"095c3d0baf601af83872c20e63e2bb7f4dd45902aba137ead9517326a397037a"} Jan 22 12:16:22 crc kubenswrapper[4874]: I0122 12:16:22.633622 4874 generic.go:334] "Generic (PLEG): container finished" podID="d598bb98-3008-4081-8187-bbd8e622bb3a" containerID="5f7d8ee5429b88b106f7fa3dbfe8f691f3605c7e687b8b5b69bea2d22edbf30c" exitCode=0 Jan 22 12:16:22 crc kubenswrapper[4874]: I0122 12:16:22.634289 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgqzt" event={"ID":"d598bb98-3008-4081-8187-bbd8e622bb3a","Type":"ContainerDied","Data":"5f7d8ee5429b88b106f7fa3dbfe8f691f3605c7e687b8b5b69bea2d22edbf30c"} Jan 22 12:16:23 crc kubenswrapper[4874]: I0122 12:16:23.653869 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgqzt" event={"ID":"d598bb98-3008-4081-8187-bbd8e622bb3a","Type":"ContainerStarted","Data":"59d9b526221eb4c75719e1f69ba088b18a62ae3391366f245f2d3a8aab22ec4e"} Jan 22 12:16:23 crc kubenswrapper[4874]: I0122 12:16:23.698136 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sgqzt" podStartSLOduration=2.266234752 podStartE2EDuration="4.698110735s" podCreationTimestamp="2026-01-22 12:16:19 +0000 UTC" firstStartedPulling="2026-01-22 12:16:20.611941165 +0000 UTC m=+2154.457012235" lastFinishedPulling="2026-01-22 12:16:23.043817138 +0000 UTC m=+2156.888888218" observedRunningTime="2026-01-22 12:16:23.693789252 +0000 UTC m=+2157.538860322" watchObservedRunningTime="2026-01-22 12:16:23.698110735 +0000 UTC m=+2157.543181815" Jan 22 12:16:29 crc kubenswrapper[4874]: I0122 12:16:29.536466 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:29 crc kubenswrapper[4874]: I0122 12:16:29.537202 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:29 crc kubenswrapper[4874]: I0122 12:16:29.598637 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:29 crc kubenswrapper[4874]: I0122 12:16:29.775376 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:31 crc kubenswrapper[4874]: I0122 12:16:31.973588 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sgqzt"] Jan 22 12:16:31 crc kubenswrapper[4874]: I0122 12:16:31.974172 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sgqzt" podUID="d598bb98-3008-4081-8187-bbd8e622bb3a" containerName="registry-server" containerID="cri-o://59d9b526221eb4c75719e1f69ba088b18a62ae3391366f245f2d3a8aab22ec4e" gracePeriod=2 Jan 22 12:16:32 crc kubenswrapper[4874]: I0122 12:16:32.756963 4874 generic.go:334] "Generic (PLEG): container finished" podID="d598bb98-3008-4081-8187-bbd8e622bb3a" containerID="59d9b526221eb4c75719e1f69ba088b18a62ae3391366f245f2d3a8aab22ec4e" exitCode=0 Jan 22 12:16:32 crc kubenswrapper[4874]: I0122 12:16:32.757251 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgqzt" event={"ID":"d598bb98-3008-4081-8187-bbd8e622bb3a","Type":"ContainerDied","Data":"59d9b526221eb4c75719e1f69ba088b18a62ae3391366f245f2d3a8aab22ec4e"} Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.558633 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.675641 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-utilities\") pod \"d598bb98-3008-4081-8187-bbd8e622bb3a\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.675753 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzgsk\" (UniqueName: \"kubernetes.io/projected/d598bb98-3008-4081-8187-bbd8e622bb3a-kube-api-access-xzgsk\") pod \"d598bb98-3008-4081-8187-bbd8e622bb3a\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.676065 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-catalog-content\") pod \"d598bb98-3008-4081-8187-bbd8e622bb3a\" (UID: \"d598bb98-3008-4081-8187-bbd8e622bb3a\") " Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.677433 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-utilities" (OuterVolumeSpecName: "utilities") pod "d598bb98-3008-4081-8187-bbd8e622bb3a" (UID: "d598bb98-3008-4081-8187-bbd8e622bb3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.704975 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d598bb98-3008-4081-8187-bbd8e622bb3a-kube-api-access-xzgsk" (OuterVolumeSpecName: "kube-api-access-xzgsk") pod "d598bb98-3008-4081-8187-bbd8e622bb3a" (UID: "d598bb98-3008-4081-8187-bbd8e622bb3a"). InnerVolumeSpecName "kube-api-access-xzgsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.740531 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d598bb98-3008-4081-8187-bbd8e622bb3a" (UID: "d598bb98-3008-4081-8187-bbd8e622bb3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.770534 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgqzt" event={"ID":"d598bb98-3008-4081-8187-bbd8e622bb3a","Type":"ContainerDied","Data":"095c3d0baf601af83872c20e63e2bb7f4dd45902aba137ead9517326a397037a"} Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.770589 4874 scope.go:117] "RemoveContainer" containerID="59d9b526221eb4c75719e1f69ba088b18a62ae3391366f245f2d3a8aab22ec4e" Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.770759 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgqzt" Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.778673 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.778721 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d598bb98-3008-4081-8187-bbd8e622bb3a-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.778742 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzgsk\" (UniqueName: \"kubernetes.io/projected/d598bb98-3008-4081-8187-bbd8e622bb3a-kube-api-access-xzgsk\") on node \"crc\" DevicePath \"\"" Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.788498 4874 scope.go:117] "RemoveContainer" containerID="5f7d8ee5429b88b106f7fa3dbfe8f691f3605c7e687b8b5b69bea2d22edbf30c" Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.809744 4874 scope.go:117] "RemoveContainer" containerID="31aa3e5fdae92637dce6ea824242e7b91959d9977ef5dd71d07c2a22cad8d5de" Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.873152 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sgqzt"] Jan 22 12:16:33 crc kubenswrapper[4874]: I0122 12:16:33.876937 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sgqzt"] Jan 22 12:16:34 crc kubenswrapper[4874]: I0122 12:16:34.729893 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d598bb98-3008-4081-8187-bbd8e622bb3a" path="/var/lib/kubelet/pods/d598bb98-3008-4081-8187-bbd8e622bb3a/volumes" Jan 22 12:17:13 crc kubenswrapper[4874]: I0122 12:17:13.520559 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:17:13 crc kubenswrapper[4874]: I0122 12:17:13.521174 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:17:43 crc kubenswrapper[4874]: I0122 12:17:43.521083 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:17:43 crc kubenswrapper[4874]: I0122 12:17:43.521757 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:18:13 crc kubenswrapper[4874]: I0122 12:18:13.520342 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:18:13 crc kubenswrapper[4874]: I0122 12:18:13.521040 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:18:13 crc kubenswrapper[4874]: I0122 12:18:13.521092 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 12:18:13 crc kubenswrapper[4874]: I0122 12:18:13.521748 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 12:18:13 crc kubenswrapper[4874]: I0122 12:18:13.521815 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" gracePeriod=600 Jan 22 12:18:13 crc kubenswrapper[4874]: E0122 12:18:13.653738 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:18:13 crc kubenswrapper[4874]: I0122 12:18:13.707523 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" exitCode=0 Jan 22 12:18:13 crc kubenswrapper[4874]: I0122 12:18:13.707581 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646"} Jan 22 12:18:13 crc kubenswrapper[4874]: I0122 12:18:13.707628 4874 scope.go:117] "RemoveContainer" containerID="4bd591f21c90200802db0c3a40e96e2477cc7ef98ac950449bf351a75f78ca92" Jan 22 12:18:13 crc kubenswrapper[4874]: I0122 12:18:13.710588 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:18:13 crc kubenswrapper[4874]: E0122 12:18:13.711095 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:18:27 crc kubenswrapper[4874]: I0122 12:18:27.716072 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:18:27 crc kubenswrapper[4874]: E0122 12:18:27.717285 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:18:39 crc kubenswrapper[4874]: I0122 12:18:39.717164 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:18:39 crc kubenswrapper[4874]: E0122 12:18:39.718512 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:18:54 crc kubenswrapper[4874]: I0122 12:18:54.723178 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:18:54 crc kubenswrapper[4874]: E0122 12:18:54.724208 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:19:06 crc kubenswrapper[4874]: I0122 12:19:06.726843 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:19:06 crc kubenswrapper[4874]: E0122 12:19:06.728099 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:19:18 crc kubenswrapper[4874]: I0122 12:19:18.716327 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:19:18 crc kubenswrapper[4874]: E0122 12:19:18.717504 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:19:29 crc kubenswrapper[4874]: I0122 12:19:29.716531 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:19:29 crc kubenswrapper[4874]: E0122 12:19:29.717242 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:19:42 crc kubenswrapper[4874]: I0122 12:19:42.716512 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:19:42 crc kubenswrapper[4874]: E0122 12:19:42.717484 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:19:57 crc kubenswrapper[4874]: I0122 12:19:57.716312 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:19:57 crc kubenswrapper[4874]: E0122 12:19:57.717466 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:20:10 crc kubenswrapper[4874]: I0122 12:20:10.716298 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:20:10 crc kubenswrapper[4874]: E0122 12:20:10.717596 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:20:21 crc kubenswrapper[4874]: I0122 12:20:21.718199 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:20:21 crc kubenswrapper[4874]: E0122 12:20:21.718885 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:20:34 crc kubenswrapper[4874]: I0122 12:20:34.716230 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:20:34 crc kubenswrapper[4874]: E0122 12:20:34.717119 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:20:46 crc kubenswrapper[4874]: I0122 12:20:46.724653 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:20:46 crc kubenswrapper[4874]: E0122 12:20:46.725667 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:20:59 crc kubenswrapper[4874]: I0122 12:20:59.716576 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:20:59 crc kubenswrapper[4874]: E0122 12:20:59.719392 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:21:12 crc kubenswrapper[4874]: I0122 12:21:12.716830 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:21:12 crc kubenswrapper[4874]: E0122 12:21:12.717763 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:21:27 crc kubenswrapper[4874]: I0122 12:21:27.716296 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:21:27 crc kubenswrapper[4874]: E0122 12:21:27.718653 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:21:40 crc kubenswrapper[4874]: I0122 12:21:40.716283 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:21:40 crc kubenswrapper[4874]: E0122 12:21:40.718077 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:21:51 crc kubenswrapper[4874]: I0122 12:21:51.715934 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:21:51 crc kubenswrapper[4874]: E0122 12:21:51.716669 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:22:02 crc kubenswrapper[4874]: I0122 12:22:02.717289 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:22:02 crc kubenswrapper[4874]: E0122 12:22:02.718310 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:22:15 crc kubenswrapper[4874]: I0122 12:22:15.716273 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:22:15 crc kubenswrapper[4874]: E0122 12:22:15.716925 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:22:28 crc kubenswrapper[4874]: I0122 12:22:28.716825 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:22:28 crc kubenswrapper[4874]: E0122 12:22:28.717743 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:22:39 crc kubenswrapper[4874]: I0122 12:22:39.720933 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:22:39 crc kubenswrapper[4874]: E0122 12:22:39.721789 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:22:52 crc kubenswrapper[4874]: I0122 12:22:52.716661 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:22:52 crc kubenswrapper[4874]: E0122 12:22:52.717596 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:22:53 crc kubenswrapper[4874]: I0122 12:22:53.874279 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wmx9h"] Jan 22 12:22:53 crc kubenswrapper[4874]: E0122 12:22:53.875244 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d598bb98-3008-4081-8187-bbd8e622bb3a" containerName="extract-utilities" Jan 22 12:22:53 crc kubenswrapper[4874]: I0122 12:22:53.875279 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d598bb98-3008-4081-8187-bbd8e622bb3a" containerName="extract-utilities" Jan 22 12:22:53 crc kubenswrapper[4874]: E0122 12:22:53.875302 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d598bb98-3008-4081-8187-bbd8e622bb3a" containerName="extract-content" Jan 22 12:22:53 crc kubenswrapper[4874]: I0122 12:22:53.875320 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d598bb98-3008-4081-8187-bbd8e622bb3a" containerName="extract-content" Jan 22 12:22:53 crc kubenswrapper[4874]: E0122 12:22:53.875383 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d598bb98-3008-4081-8187-bbd8e622bb3a" containerName="registry-server" Jan 22 12:22:53 crc kubenswrapper[4874]: I0122 12:22:53.875428 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d598bb98-3008-4081-8187-bbd8e622bb3a" containerName="registry-server" Jan 22 12:22:53 crc kubenswrapper[4874]: I0122 12:22:53.875764 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="d598bb98-3008-4081-8187-bbd8e622bb3a" containerName="registry-server" Jan 22 12:22:53 crc kubenswrapper[4874]: I0122 12:22:53.878050 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:22:53 crc kubenswrapper[4874]: I0122 12:22:53.904649 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wmx9h"] Jan 22 12:22:53 crc kubenswrapper[4874]: I0122 12:22:53.968469 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wjjg\" (UniqueName: \"kubernetes.io/projected/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-kube-api-access-6wjjg\") pod \"certified-operators-wmx9h\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:22:53 crc kubenswrapper[4874]: I0122 12:22:53.968604 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-utilities\") pod \"certified-operators-wmx9h\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:22:53 crc kubenswrapper[4874]: I0122 12:22:53.968731 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-catalog-content\") pod \"certified-operators-wmx9h\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:22:54 crc kubenswrapper[4874]: I0122 12:22:54.070512 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wjjg\" (UniqueName: \"kubernetes.io/projected/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-kube-api-access-6wjjg\") pod \"certified-operators-wmx9h\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:22:54 crc kubenswrapper[4874]: I0122 12:22:54.070579 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-utilities\") pod \"certified-operators-wmx9h\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:22:54 crc kubenswrapper[4874]: I0122 12:22:54.070654 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-catalog-content\") pod \"certified-operators-wmx9h\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:22:54 crc kubenswrapper[4874]: I0122 12:22:54.071351 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-catalog-content\") pod \"certified-operators-wmx9h\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:22:54 crc kubenswrapper[4874]: I0122 12:22:54.072172 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-utilities\") pod \"certified-operators-wmx9h\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:22:54 crc kubenswrapper[4874]: I0122 12:22:54.096850 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wjjg\" (UniqueName: \"kubernetes.io/projected/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-kube-api-access-6wjjg\") pod \"certified-operators-wmx9h\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:22:54 crc kubenswrapper[4874]: I0122 12:22:54.199488 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:22:54 crc kubenswrapper[4874]: I0122 12:22:54.495949 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wmx9h"] Jan 22 12:22:54 crc kubenswrapper[4874]: I0122 12:22:54.998271 4874 generic.go:334] "Generic (PLEG): container finished" podID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" containerID="947367f2e70a6cdb0c481406ee1ca0c3c4d739abf78030cfbf933e595b9ca109" exitCode=0 Jan 22 12:22:54 crc kubenswrapper[4874]: I0122 12:22:54.998324 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmx9h" event={"ID":"d6f3e52c-c484-443a-aee8-bfc31e28c9b8","Type":"ContainerDied","Data":"947367f2e70a6cdb0c481406ee1ca0c3c4d739abf78030cfbf933e595b9ca109"} Jan 22 12:22:54 crc kubenswrapper[4874]: I0122 12:22:54.998359 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmx9h" event={"ID":"d6f3e52c-c484-443a-aee8-bfc31e28c9b8","Type":"ContainerStarted","Data":"e43a9df2d62e043104820584dc48aa02e31ee1e3d15d3acd5111300752598cd9"} Jan 22 12:22:55 crc kubenswrapper[4874]: I0122 12:22:55.000802 4874 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 12:22:56 crc kubenswrapper[4874]: I0122 12:22:56.011624 4874 generic.go:334] "Generic (PLEG): container finished" podID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" containerID="33a21e73f13e2c4c680d2d6e9593fb5a6b498dcdb3dfac7b50ef86c6cca2642f" exitCode=0 Jan 22 12:22:56 crc kubenswrapper[4874]: I0122 12:22:56.011746 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmx9h" event={"ID":"d6f3e52c-c484-443a-aee8-bfc31e28c9b8","Type":"ContainerDied","Data":"33a21e73f13e2c4c680d2d6e9593fb5a6b498dcdb3dfac7b50ef86c6cca2642f"} Jan 22 12:22:57 crc kubenswrapper[4874]: I0122 12:22:57.023063 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmx9h" event={"ID":"d6f3e52c-c484-443a-aee8-bfc31e28c9b8","Type":"ContainerStarted","Data":"2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe"} Jan 22 12:22:57 crc kubenswrapper[4874]: I0122 12:22:57.051440 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wmx9h" podStartSLOduration=2.632070432 podStartE2EDuration="4.051415472s" podCreationTimestamp="2026-01-22 12:22:53 +0000 UTC" firstStartedPulling="2026-01-22 12:22:55.000286607 +0000 UTC m=+2548.845357717" lastFinishedPulling="2026-01-22 12:22:56.419631647 +0000 UTC m=+2550.264702757" observedRunningTime="2026-01-22 12:22:57.048201762 +0000 UTC m=+2550.893272852" watchObservedRunningTime="2026-01-22 12:22:57.051415472 +0000 UTC m=+2550.896486562" Jan 22 12:23:04 crc kubenswrapper[4874]: I0122 12:23:04.200591 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:23:04 crc kubenswrapper[4874]: I0122 12:23:04.201456 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:23:04 crc kubenswrapper[4874]: I0122 12:23:04.274598 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:23:05 crc kubenswrapper[4874]: I0122 12:23:05.166672 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:23:05 crc kubenswrapper[4874]: I0122 12:23:05.270527 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wmx9h"] Jan 22 12:23:07 crc kubenswrapper[4874]: I0122 12:23:07.121718 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wmx9h" podUID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" containerName="registry-server" containerID="cri-o://2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe" gracePeriod=2 Jan 22 12:23:07 crc kubenswrapper[4874]: I0122 12:23:07.716890 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:23:07 crc kubenswrapper[4874]: E0122 12:23:07.717826 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.092836 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.105933 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-utilities\") pod \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.106233 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wjjg\" (UniqueName: \"kubernetes.io/projected/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-kube-api-access-6wjjg\") pod \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.106281 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-catalog-content\") pod \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\" (UID: \"d6f3e52c-c484-443a-aee8-bfc31e28c9b8\") " Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.107242 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-utilities" (OuterVolumeSpecName: "utilities") pod "d6f3e52c-c484-443a-aee8-bfc31e28c9b8" (UID: "d6f3e52c-c484-443a-aee8-bfc31e28c9b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.116126 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-kube-api-access-6wjjg" (OuterVolumeSpecName: "kube-api-access-6wjjg") pod "d6f3e52c-c484-443a-aee8-bfc31e28c9b8" (UID: "d6f3e52c-c484-443a-aee8-bfc31e28c9b8"). InnerVolumeSpecName "kube-api-access-6wjjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.135236 4874 generic.go:334] "Generic (PLEG): container finished" podID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" containerID="2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe" exitCode=0 Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.135284 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmx9h" event={"ID":"d6f3e52c-c484-443a-aee8-bfc31e28c9b8","Type":"ContainerDied","Data":"2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe"} Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.135313 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmx9h" event={"ID":"d6f3e52c-c484-443a-aee8-bfc31e28c9b8","Type":"ContainerDied","Data":"e43a9df2d62e043104820584dc48aa02e31ee1e3d15d3acd5111300752598cd9"} Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.135330 4874 scope.go:117] "RemoveContainer" containerID="2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.135487 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmx9h" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.160978 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6f3e52c-c484-443a-aee8-bfc31e28c9b8" (UID: "d6f3e52c-c484-443a-aee8-bfc31e28c9b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.168209 4874 scope.go:117] "RemoveContainer" containerID="33a21e73f13e2c4c680d2d6e9593fb5a6b498dcdb3dfac7b50ef86c6cca2642f" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.186239 4874 scope.go:117] "RemoveContainer" containerID="947367f2e70a6cdb0c481406ee1ca0c3c4d739abf78030cfbf933e595b9ca109" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.208790 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wjjg\" (UniqueName: \"kubernetes.io/projected/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-kube-api-access-6wjjg\") on node \"crc\" DevicePath \"\"" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.208859 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.208896 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f3e52c-c484-443a-aee8-bfc31e28c9b8-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.223315 4874 scope.go:117] "RemoveContainer" containerID="2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe" Jan 22 12:23:08 crc kubenswrapper[4874]: E0122 12:23:08.223860 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe\": container with ID starting with 2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe not found: ID does not exist" containerID="2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.223962 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe"} err="failed to get container status \"2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe\": rpc error: code = NotFound desc = could not find container \"2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe\": container with ID starting with 2c7a2b76b6152fe45c04a7b90cc5de979ff0bf8462fcdee1fcc8edc296fadafe not found: ID does not exist" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.224046 4874 scope.go:117] "RemoveContainer" containerID="33a21e73f13e2c4c680d2d6e9593fb5a6b498dcdb3dfac7b50ef86c6cca2642f" Jan 22 12:23:08 crc kubenswrapper[4874]: E0122 12:23:08.224586 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33a21e73f13e2c4c680d2d6e9593fb5a6b498dcdb3dfac7b50ef86c6cca2642f\": container with ID starting with 33a21e73f13e2c4c680d2d6e9593fb5a6b498dcdb3dfac7b50ef86c6cca2642f not found: ID does not exist" containerID="33a21e73f13e2c4c680d2d6e9593fb5a6b498dcdb3dfac7b50ef86c6cca2642f" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.224648 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33a21e73f13e2c4c680d2d6e9593fb5a6b498dcdb3dfac7b50ef86c6cca2642f"} err="failed to get container status \"33a21e73f13e2c4c680d2d6e9593fb5a6b498dcdb3dfac7b50ef86c6cca2642f\": rpc error: code = NotFound desc = could not find container \"33a21e73f13e2c4c680d2d6e9593fb5a6b498dcdb3dfac7b50ef86c6cca2642f\": container with ID starting with 33a21e73f13e2c4c680d2d6e9593fb5a6b498dcdb3dfac7b50ef86c6cca2642f not found: ID does not exist" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.224696 4874 scope.go:117] "RemoveContainer" containerID="947367f2e70a6cdb0c481406ee1ca0c3c4d739abf78030cfbf933e595b9ca109" Jan 22 12:23:08 crc kubenswrapper[4874]: E0122 12:23:08.225091 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"947367f2e70a6cdb0c481406ee1ca0c3c4d739abf78030cfbf933e595b9ca109\": container with ID starting with 947367f2e70a6cdb0c481406ee1ca0c3c4d739abf78030cfbf933e595b9ca109 not found: ID does not exist" containerID="947367f2e70a6cdb0c481406ee1ca0c3c4d739abf78030cfbf933e595b9ca109" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.225172 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"947367f2e70a6cdb0c481406ee1ca0c3c4d739abf78030cfbf933e595b9ca109"} err="failed to get container status \"947367f2e70a6cdb0c481406ee1ca0c3c4d739abf78030cfbf933e595b9ca109\": rpc error: code = NotFound desc = could not find container \"947367f2e70a6cdb0c481406ee1ca0c3c4d739abf78030cfbf933e595b9ca109\": container with ID starting with 947367f2e70a6cdb0c481406ee1ca0c3c4d739abf78030cfbf933e595b9ca109 not found: ID does not exist" Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.503671 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wmx9h"] Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.514062 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wmx9h"] Jan 22 12:23:08 crc kubenswrapper[4874]: I0122 12:23:08.728589 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" path="/var/lib/kubelet/pods/d6f3e52c-c484-443a-aee8-bfc31e28c9b8/volumes" Jan 22 12:23:22 crc kubenswrapper[4874]: I0122 12:23:22.716516 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:23:23 crc kubenswrapper[4874]: I0122 12:23:23.280418 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"63365a511b966484f39bb84f505e4f8a844bd95a40ffac8357b30eaa5bacd904"} Jan 22 12:25:43 crc kubenswrapper[4874]: I0122 12:25:43.520146 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:25:43 crc kubenswrapper[4874]: I0122 12:25:43.520802 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:25:57 crc kubenswrapper[4874]: I0122 12:25:57.906479 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cdztk"] Jan 22 12:25:57 crc kubenswrapper[4874]: E0122 12:25:57.907671 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" containerName="registry-server" Jan 22 12:25:57 crc kubenswrapper[4874]: I0122 12:25:57.907704 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" containerName="registry-server" Jan 22 12:25:57 crc kubenswrapper[4874]: E0122 12:25:57.907730 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" containerName="extract-utilities" Jan 22 12:25:57 crc kubenswrapper[4874]: I0122 12:25:57.907747 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" containerName="extract-utilities" Jan 22 12:25:57 crc kubenswrapper[4874]: E0122 12:25:57.907793 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" containerName="extract-content" Jan 22 12:25:57 crc kubenswrapper[4874]: I0122 12:25:57.907812 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" containerName="extract-content" Jan 22 12:25:57 crc kubenswrapper[4874]: I0122 12:25:57.908058 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6f3e52c-c484-443a-aee8-bfc31e28c9b8" containerName="registry-server" Jan 22 12:25:57 crc kubenswrapper[4874]: I0122 12:25:57.910008 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:25:57 crc kubenswrapper[4874]: I0122 12:25:57.915202 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cdztk"] Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.002090 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-utilities\") pod \"redhat-operators-cdztk\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.002230 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-catalog-content\") pod \"redhat-operators-cdztk\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.002279 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxrrm\" (UniqueName: \"kubernetes.io/projected/b6dd15bf-729b-415a-835f-20a2d777fec0-kube-api-access-dxrrm\") pod \"redhat-operators-cdztk\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.103110 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-utilities\") pod \"redhat-operators-cdztk\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.103244 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-catalog-content\") pod \"redhat-operators-cdztk\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.103289 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxrrm\" (UniqueName: \"kubernetes.io/projected/b6dd15bf-729b-415a-835f-20a2d777fec0-kube-api-access-dxrrm\") pod \"redhat-operators-cdztk\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.103568 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-utilities\") pod \"redhat-operators-cdztk\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.103699 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-catalog-content\") pod \"redhat-operators-cdztk\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.125357 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxrrm\" (UniqueName: \"kubernetes.io/projected/b6dd15bf-729b-415a-835f-20a2d777fec0-kube-api-access-dxrrm\") pod \"redhat-operators-cdztk\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.238967 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.518695 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cdztk"] Jan 22 12:25:58 crc kubenswrapper[4874]: I0122 12:25:58.669289 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdztk" event={"ID":"b6dd15bf-729b-415a-835f-20a2d777fec0","Type":"ContainerStarted","Data":"03ec6cb9b7fadd34d1fe75156c0aa5d424f939eecdfd48ef6b48f534c40ec2e0"} Jan 22 12:25:59 crc kubenswrapper[4874]: I0122 12:25:59.677928 4874 generic.go:334] "Generic (PLEG): container finished" podID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerID="3256cbd3318e852d8cc4290f2bb70ae12b509bf1d1903df37059cd9dca3f3ac5" exitCode=0 Jan 22 12:25:59 crc kubenswrapper[4874]: I0122 12:25:59.678039 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdztk" event={"ID":"b6dd15bf-729b-415a-835f-20a2d777fec0","Type":"ContainerDied","Data":"3256cbd3318e852d8cc4290f2bb70ae12b509bf1d1903df37059cd9dca3f3ac5"} Jan 22 12:26:00 crc kubenswrapper[4874]: I0122 12:26:00.688001 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdztk" event={"ID":"b6dd15bf-729b-415a-835f-20a2d777fec0","Type":"ContainerStarted","Data":"22314f87718ae2d216abf1d85d2d6624b190fa8642a36d93ae61ae2b86956b68"} Jan 22 12:26:01 crc kubenswrapper[4874]: I0122 12:26:01.699961 4874 generic.go:334] "Generic (PLEG): container finished" podID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerID="22314f87718ae2d216abf1d85d2d6624b190fa8642a36d93ae61ae2b86956b68" exitCode=0 Jan 22 12:26:01 crc kubenswrapper[4874]: I0122 12:26:01.700031 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdztk" event={"ID":"b6dd15bf-729b-415a-835f-20a2d777fec0","Type":"ContainerDied","Data":"22314f87718ae2d216abf1d85d2d6624b190fa8642a36d93ae61ae2b86956b68"} Jan 22 12:26:02 crc kubenswrapper[4874]: I0122 12:26:02.730731 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdztk" event={"ID":"b6dd15bf-729b-415a-835f-20a2d777fec0","Type":"ContainerStarted","Data":"d95550138a7512b811fbb08caadebf233ff8761506de6ed159f6c97c657a2cb4"} Jan 22 12:26:02 crc kubenswrapper[4874]: I0122 12:26:02.745956 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cdztk" podStartSLOduration=3.098073188 podStartE2EDuration="5.745936145s" podCreationTimestamp="2026-01-22 12:25:57 +0000 UTC" firstStartedPulling="2026-01-22 12:25:59.679882421 +0000 UTC m=+2733.524953491" lastFinishedPulling="2026-01-22 12:26:02.327745338 +0000 UTC m=+2736.172816448" observedRunningTime="2026-01-22 12:26:02.736536505 +0000 UTC m=+2736.581607585" watchObservedRunningTime="2026-01-22 12:26:02.745936145 +0000 UTC m=+2736.591007225" Jan 22 12:26:08 crc kubenswrapper[4874]: I0122 12:26:08.239600 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:26:08 crc kubenswrapper[4874]: I0122 12:26:08.240689 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:26:09 crc kubenswrapper[4874]: I0122 12:26:09.290239 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cdztk" podUID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerName="registry-server" probeResult="failure" output=< Jan 22 12:26:09 crc kubenswrapper[4874]: timeout: failed to connect service ":50051" within 1s Jan 22 12:26:09 crc kubenswrapper[4874]: > Jan 22 12:26:13 crc kubenswrapper[4874]: I0122 12:26:13.520974 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:26:13 crc kubenswrapper[4874]: I0122 12:26:13.521387 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:26:18 crc kubenswrapper[4874]: I0122 12:26:18.283621 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:26:18 crc kubenswrapper[4874]: I0122 12:26:18.343379 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:26:19 crc kubenswrapper[4874]: I0122 12:26:19.801763 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cdztk"] Jan 22 12:26:19 crc kubenswrapper[4874]: I0122 12:26:19.862375 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cdztk" podUID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerName="registry-server" containerID="cri-o://d95550138a7512b811fbb08caadebf233ff8761506de6ed159f6c97c657a2cb4" gracePeriod=2 Jan 22 12:26:20 crc kubenswrapper[4874]: I0122 12:26:20.877446 4874 generic.go:334] "Generic (PLEG): container finished" podID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerID="d95550138a7512b811fbb08caadebf233ff8761506de6ed159f6c97c657a2cb4" exitCode=0 Jan 22 12:26:20 crc kubenswrapper[4874]: I0122 12:26:20.877498 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdztk" event={"ID":"b6dd15bf-729b-415a-835f-20a2d777fec0","Type":"ContainerDied","Data":"d95550138a7512b811fbb08caadebf233ff8761506de6ed159f6c97c657a2cb4"} Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.379894 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.535120 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-catalog-content\") pod \"b6dd15bf-729b-415a-835f-20a2d777fec0\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.535182 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxrrm\" (UniqueName: \"kubernetes.io/projected/b6dd15bf-729b-415a-835f-20a2d777fec0-kube-api-access-dxrrm\") pod \"b6dd15bf-729b-415a-835f-20a2d777fec0\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.535263 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-utilities\") pod \"b6dd15bf-729b-415a-835f-20a2d777fec0\" (UID: \"b6dd15bf-729b-415a-835f-20a2d777fec0\") " Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.536428 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-utilities" (OuterVolumeSpecName: "utilities") pod "b6dd15bf-729b-415a-835f-20a2d777fec0" (UID: "b6dd15bf-729b-415a-835f-20a2d777fec0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.552890 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6dd15bf-729b-415a-835f-20a2d777fec0-kube-api-access-dxrrm" (OuterVolumeSpecName: "kube-api-access-dxrrm") pod "b6dd15bf-729b-415a-835f-20a2d777fec0" (UID: "b6dd15bf-729b-415a-835f-20a2d777fec0"). InnerVolumeSpecName "kube-api-access-dxrrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.637092 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxrrm\" (UniqueName: \"kubernetes.io/projected/b6dd15bf-729b-415a-835f-20a2d777fec0-kube-api-access-dxrrm\") on node \"crc\" DevicePath \"\"" Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.637141 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.685533 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6dd15bf-729b-415a-835f-20a2d777fec0" (UID: "b6dd15bf-729b-415a-835f-20a2d777fec0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.740547 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6dd15bf-729b-415a-835f-20a2d777fec0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.888374 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdztk" event={"ID":"b6dd15bf-729b-415a-835f-20a2d777fec0","Type":"ContainerDied","Data":"03ec6cb9b7fadd34d1fe75156c0aa5d424f939eecdfd48ef6b48f534c40ec2e0"} Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.888453 4874 scope.go:117] "RemoveContainer" containerID="d95550138a7512b811fbb08caadebf233ff8761506de6ed159f6c97c657a2cb4" Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.888495 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cdztk" Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.913696 4874 scope.go:117] "RemoveContainer" containerID="22314f87718ae2d216abf1d85d2d6624b190fa8642a36d93ae61ae2b86956b68" Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.935770 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cdztk"] Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.947265 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cdztk"] Jan 22 12:26:21 crc kubenswrapper[4874]: I0122 12:26:21.951697 4874 scope.go:117] "RemoveContainer" containerID="3256cbd3318e852d8cc4290f2bb70ae12b509bf1d1903df37059cd9dca3f3ac5" Jan 22 12:26:22 crc kubenswrapper[4874]: I0122 12:26:22.726078 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6dd15bf-729b-415a-835f-20a2d777fec0" path="/var/lib/kubelet/pods/b6dd15bf-729b-415a-835f-20a2d777fec0/volumes" Jan 22 12:26:43 crc kubenswrapper[4874]: I0122 12:26:43.520016 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:26:43 crc kubenswrapper[4874]: I0122 12:26:43.520531 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:26:43 crc kubenswrapper[4874]: I0122 12:26:43.520581 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 12:26:43 crc kubenswrapper[4874]: I0122 12:26:43.521060 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"63365a511b966484f39bb84f505e4f8a844bd95a40ffac8357b30eaa5bacd904"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 12:26:43 crc kubenswrapper[4874]: I0122 12:26:43.521113 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://63365a511b966484f39bb84f505e4f8a844bd95a40ffac8357b30eaa5bacd904" gracePeriod=600 Jan 22 12:26:44 crc kubenswrapper[4874]: I0122 12:26:44.047452 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="63365a511b966484f39bb84f505e4f8a844bd95a40ffac8357b30eaa5bacd904" exitCode=0 Jan 22 12:26:44 crc kubenswrapper[4874]: I0122 12:26:44.047605 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"63365a511b966484f39bb84f505e4f8a844bd95a40ffac8357b30eaa5bacd904"} Jan 22 12:26:44 crc kubenswrapper[4874]: I0122 12:26:44.047831 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7"} Jan 22 12:26:44 crc kubenswrapper[4874]: I0122 12:26:44.047851 4874 scope.go:117] "RemoveContainer" containerID="b29ca28e4593590c27126c3de6633dd0a7e93ba84ad80eaa6d8fb29df1080646" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.073758 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t99tf"] Jan 22 12:27:03 crc kubenswrapper[4874]: E0122 12:27:03.076374 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerName="extract-content" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.076424 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerName="extract-content" Jan 22 12:27:03 crc kubenswrapper[4874]: E0122 12:27:03.076472 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerName="extract-utilities" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.076485 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerName="extract-utilities" Jan 22 12:27:03 crc kubenswrapper[4874]: E0122 12:27:03.076508 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerName="registry-server" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.076520 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerName="registry-server" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.076718 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6dd15bf-729b-415a-835f-20a2d777fec0" containerName="registry-server" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.078269 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.086416 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t99tf"] Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.246291 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6xt6\" (UniqueName: \"kubernetes.io/projected/fda9be7d-c4d8-474a-956a-20de22b93a47-kube-api-access-q6xt6\") pod \"community-operators-t99tf\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.246383 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-utilities\") pod \"community-operators-t99tf\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.246453 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-catalog-content\") pod \"community-operators-t99tf\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.347827 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-utilities\") pod \"community-operators-t99tf\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.347897 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-catalog-content\") pod \"community-operators-t99tf\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.347985 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6xt6\" (UniqueName: \"kubernetes.io/projected/fda9be7d-c4d8-474a-956a-20de22b93a47-kube-api-access-q6xt6\") pod \"community-operators-t99tf\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.348494 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-utilities\") pod \"community-operators-t99tf\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.348551 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-catalog-content\") pod \"community-operators-t99tf\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.367329 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6xt6\" (UniqueName: \"kubernetes.io/projected/fda9be7d-c4d8-474a-956a-20de22b93a47-kube-api-access-q6xt6\") pod \"community-operators-t99tf\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.453780 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:03 crc kubenswrapper[4874]: I0122 12:27:03.968886 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t99tf"] Jan 22 12:27:04 crc kubenswrapper[4874]: I0122 12:27:04.221055 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t99tf" event={"ID":"fda9be7d-c4d8-474a-956a-20de22b93a47","Type":"ContainerDied","Data":"db3c128e689667f735cf9e0a6ed9f2088a04522ca182fa45cdfc975c925fa480"} Jan 22 12:27:04 crc kubenswrapper[4874]: I0122 12:27:04.220814 4874 generic.go:334] "Generic (PLEG): container finished" podID="fda9be7d-c4d8-474a-956a-20de22b93a47" containerID="db3c128e689667f735cf9e0a6ed9f2088a04522ca182fa45cdfc975c925fa480" exitCode=0 Jan 22 12:27:04 crc kubenswrapper[4874]: I0122 12:27:04.222238 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t99tf" event={"ID":"fda9be7d-c4d8-474a-956a-20de22b93a47","Type":"ContainerStarted","Data":"607e726b85edcce4faa88a5cb68ab4a274cf859ac65ee4cb189008767f474219"} Jan 22 12:27:05 crc kubenswrapper[4874]: I0122 12:27:05.234798 4874 generic.go:334] "Generic (PLEG): container finished" podID="fda9be7d-c4d8-474a-956a-20de22b93a47" containerID="bce84a4d49a705a423820b0b8e7f8f9955c5910e4f4352f68f63e8fa8aaf2782" exitCode=0 Jan 22 12:27:05 crc kubenswrapper[4874]: I0122 12:27:05.234944 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t99tf" event={"ID":"fda9be7d-c4d8-474a-956a-20de22b93a47","Type":"ContainerDied","Data":"bce84a4d49a705a423820b0b8e7f8f9955c5910e4f4352f68f63e8fa8aaf2782"} Jan 22 12:27:06 crc kubenswrapper[4874]: I0122 12:27:06.254071 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t99tf" event={"ID":"fda9be7d-c4d8-474a-956a-20de22b93a47","Type":"ContainerStarted","Data":"78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72"} Jan 22 12:27:13 crc kubenswrapper[4874]: I0122 12:27:13.454730 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:13 crc kubenswrapper[4874]: I0122 12:27:13.456785 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:13 crc kubenswrapper[4874]: I0122 12:27:13.529610 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:13 crc kubenswrapper[4874]: I0122 12:27:13.559588 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t99tf" podStartSLOduration=9.098910049 podStartE2EDuration="10.559572866s" podCreationTimestamp="2026-01-22 12:27:03 +0000 UTC" firstStartedPulling="2026-01-22 12:27:04.222857734 +0000 UTC m=+2798.067928844" lastFinishedPulling="2026-01-22 12:27:05.683520551 +0000 UTC m=+2799.528591661" observedRunningTime="2026-01-22 12:27:06.290921733 +0000 UTC m=+2800.135992813" watchObservedRunningTime="2026-01-22 12:27:13.559572866 +0000 UTC m=+2807.404643936" Jan 22 12:27:14 crc kubenswrapper[4874]: I0122 12:27:14.364974 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:14 crc kubenswrapper[4874]: I0122 12:27:14.415102 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t99tf"] Jan 22 12:27:16 crc kubenswrapper[4874]: I0122 12:27:16.342554 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t99tf" podUID="fda9be7d-c4d8-474a-956a-20de22b93a47" containerName="registry-server" containerID="cri-o://78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72" gracePeriod=2 Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.281604 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.350804 4874 generic.go:334] "Generic (PLEG): container finished" podID="fda9be7d-c4d8-474a-956a-20de22b93a47" containerID="78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72" exitCode=0 Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.350850 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t99tf" event={"ID":"fda9be7d-c4d8-474a-956a-20de22b93a47","Type":"ContainerDied","Data":"78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72"} Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.350866 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t99tf" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.350878 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t99tf" event={"ID":"fda9be7d-c4d8-474a-956a-20de22b93a47","Type":"ContainerDied","Data":"607e726b85edcce4faa88a5cb68ab4a274cf859ac65ee4cb189008767f474219"} Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.350896 4874 scope.go:117] "RemoveContainer" containerID="78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.366993 4874 scope.go:117] "RemoveContainer" containerID="bce84a4d49a705a423820b0b8e7f8f9955c5910e4f4352f68f63e8fa8aaf2782" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.380226 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6xt6\" (UniqueName: \"kubernetes.io/projected/fda9be7d-c4d8-474a-956a-20de22b93a47-kube-api-access-q6xt6\") pod \"fda9be7d-c4d8-474a-956a-20de22b93a47\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.380279 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-catalog-content\") pod \"fda9be7d-c4d8-474a-956a-20de22b93a47\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.380321 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-utilities\") pod \"fda9be7d-c4d8-474a-956a-20de22b93a47\" (UID: \"fda9be7d-c4d8-474a-956a-20de22b93a47\") " Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.381679 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-utilities" (OuterVolumeSpecName: "utilities") pod "fda9be7d-c4d8-474a-956a-20de22b93a47" (UID: "fda9be7d-c4d8-474a-956a-20de22b93a47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.385772 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda9be7d-c4d8-474a-956a-20de22b93a47-kube-api-access-q6xt6" (OuterVolumeSpecName: "kube-api-access-q6xt6") pod "fda9be7d-c4d8-474a-956a-20de22b93a47" (UID: "fda9be7d-c4d8-474a-956a-20de22b93a47"). InnerVolumeSpecName "kube-api-access-q6xt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.391985 4874 scope.go:117] "RemoveContainer" containerID="db3c128e689667f735cf9e0a6ed9f2088a04522ca182fa45cdfc975c925fa480" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.431646 4874 scope.go:117] "RemoveContainer" containerID="78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72" Jan 22 12:27:17 crc kubenswrapper[4874]: E0122 12:27:17.432080 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72\": container with ID starting with 78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72 not found: ID does not exist" containerID="78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.432118 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72"} err="failed to get container status \"78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72\": rpc error: code = NotFound desc = could not find container \"78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72\": container with ID starting with 78cf9858463bcb9cda28f3ce4442e4717821d4e2b6fe395ff42951574324ae72 not found: ID does not exist" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.432142 4874 scope.go:117] "RemoveContainer" containerID="bce84a4d49a705a423820b0b8e7f8f9955c5910e4f4352f68f63e8fa8aaf2782" Jan 22 12:27:17 crc kubenswrapper[4874]: E0122 12:27:17.432653 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bce84a4d49a705a423820b0b8e7f8f9955c5910e4f4352f68f63e8fa8aaf2782\": container with ID starting with bce84a4d49a705a423820b0b8e7f8f9955c5910e4f4352f68f63e8fa8aaf2782 not found: ID does not exist" containerID="bce84a4d49a705a423820b0b8e7f8f9955c5910e4f4352f68f63e8fa8aaf2782" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.432676 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bce84a4d49a705a423820b0b8e7f8f9955c5910e4f4352f68f63e8fa8aaf2782"} err="failed to get container status \"bce84a4d49a705a423820b0b8e7f8f9955c5910e4f4352f68f63e8fa8aaf2782\": rpc error: code = NotFound desc = could not find container \"bce84a4d49a705a423820b0b8e7f8f9955c5910e4f4352f68f63e8fa8aaf2782\": container with ID starting with bce84a4d49a705a423820b0b8e7f8f9955c5910e4f4352f68f63e8fa8aaf2782 not found: ID does not exist" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.432693 4874 scope.go:117] "RemoveContainer" containerID="db3c128e689667f735cf9e0a6ed9f2088a04522ca182fa45cdfc975c925fa480" Jan 22 12:27:17 crc kubenswrapper[4874]: E0122 12:27:17.432988 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db3c128e689667f735cf9e0a6ed9f2088a04522ca182fa45cdfc975c925fa480\": container with ID starting with db3c128e689667f735cf9e0a6ed9f2088a04522ca182fa45cdfc975c925fa480 not found: ID does not exist" containerID="db3c128e689667f735cf9e0a6ed9f2088a04522ca182fa45cdfc975c925fa480" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.433009 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db3c128e689667f735cf9e0a6ed9f2088a04522ca182fa45cdfc975c925fa480"} err="failed to get container status \"db3c128e689667f735cf9e0a6ed9f2088a04522ca182fa45cdfc975c925fa480\": rpc error: code = NotFound desc = could not find container \"db3c128e689667f735cf9e0a6ed9f2088a04522ca182fa45cdfc975c925fa480\": container with ID starting with db3c128e689667f735cf9e0a6ed9f2088a04522ca182fa45cdfc975c925fa480 not found: ID does not exist" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.438335 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fda9be7d-c4d8-474a-956a-20de22b93a47" (UID: "fda9be7d-c4d8-474a-956a-20de22b93a47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.481736 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6xt6\" (UniqueName: \"kubernetes.io/projected/fda9be7d-c4d8-474a-956a-20de22b93a47-kube-api-access-q6xt6\") on node \"crc\" DevicePath \"\"" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.481771 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.481784 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fda9be7d-c4d8-474a-956a-20de22b93a47-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.700417 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t99tf"] Jan 22 12:27:17 crc kubenswrapper[4874]: I0122 12:27:17.708907 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t99tf"] Jan 22 12:27:18 crc kubenswrapper[4874]: I0122 12:27:18.743604 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda9be7d-c4d8-474a-956a-20de22b93a47" path="/var/lib/kubelet/pods/fda9be7d-c4d8-474a-956a-20de22b93a47/volumes" Jan 22 12:28:43 crc kubenswrapper[4874]: I0122 12:28:43.520764 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:28:43 crc kubenswrapper[4874]: I0122 12:28:43.521430 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:29:13 crc kubenswrapper[4874]: I0122 12:29:13.520919 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:29:13 crc kubenswrapper[4874]: I0122 12:29:13.521595 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:29:43 crc kubenswrapper[4874]: I0122 12:29:43.520846 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:29:43 crc kubenswrapper[4874]: I0122 12:29:43.521498 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:29:43 crc kubenswrapper[4874]: I0122 12:29:43.521574 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 12:29:43 crc kubenswrapper[4874]: I0122 12:29:43.522587 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 12:29:43 crc kubenswrapper[4874]: I0122 12:29:43.522707 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" gracePeriod=600 Jan 22 12:29:43 crc kubenswrapper[4874]: E0122 12:29:43.669149 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:29:43 crc kubenswrapper[4874]: I0122 12:29:43.675026 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" exitCode=0 Jan 22 12:29:43 crc kubenswrapper[4874]: I0122 12:29:43.675100 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7"} Jan 22 12:29:43 crc kubenswrapper[4874]: I0122 12:29:43.675154 4874 scope.go:117] "RemoveContainer" containerID="63365a511b966484f39bb84f505e4f8a844bd95a40ffac8357b30eaa5bacd904" Jan 22 12:29:43 crc kubenswrapper[4874]: I0122 12:29:43.675980 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:29:43 crc kubenswrapper[4874]: E0122 12:29:43.677156 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:29:56 crc kubenswrapper[4874]: I0122 12:29:56.724273 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:29:56 crc kubenswrapper[4874]: E0122 12:29:56.725326 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.158722 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7"] Jan 22 12:30:00 crc kubenswrapper[4874]: E0122 12:30:00.159499 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fda9be7d-c4d8-474a-956a-20de22b93a47" containerName="registry-server" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.159513 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="fda9be7d-c4d8-474a-956a-20de22b93a47" containerName="registry-server" Jan 22 12:30:00 crc kubenswrapper[4874]: E0122 12:30:00.159531 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fda9be7d-c4d8-474a-956a-20de22b93a47" containerName="extract-utilities" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.159540 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="fda9be7d-c4d8-474a-956a-20de22b93a47" containerName="extract-utilities" Jan 22 12:30:00 crc kubenswrapper[4874]: E0122 12:30:00.159560 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fda9be7d-c4d8-474a-956a-20de22b93a47" containerName="extract-content" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.159568 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="fda9be7d-c4d8-474a-956a-20de22b93a47" containerName="extract-content" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.159731 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="fda9be7d-c4d8-474a-956a-20de22b93a47" containerName="registry-server" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.160300 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.162747 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.167898 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.173101 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7"] Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.318341 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9a6256a-adf8-4648-bc08-825f3d89a607-config-volume\") pod \"collect-profiles-29484750-cfdm7\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.318541 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w7z5\" (UniqueName: \"kubernetes.io/projected/b9a6256a-adf8-4648-bc08-825f3d89a607-kube-api-access-9w7z5\") pod \"collect-profiles-29484750-cfdm7\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.318587 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9a6256a-adf8-4648-bc08-825f3d89a607-secret-volume\") pod \"collect-profiles-29484750-cfdm7\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.419841 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w7z5\" (UniqueName: \"kubernetes.io/projected/b9a6256a-adf8-4648-bc08-825f3d89a607-kube-api-access-9w7z5\") pod \"collect-profiles-29484750-cfdm7\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.419934 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9a6256a-adf8-4648-bc08-825f3d89a607-secret-volume\") pod \"collect-profiles-29484750-cfdm7\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.420036 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9a6256a-adf8-4648-bc08-825f3d89a607-config-volume\") pod \"collect-profiles-29484750-cfdm7\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.421987 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9a6256a-adf8-4648-bc08-825f3d89a607-config-volume\") pod \"collect-profiles-29484750-cfdm7\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.430547 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9a6256a-adf8-4648-bc08-825f3d89a607-secret-volume\") pod \"collect-profiles-29484750-cfdm7\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.444559 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w7z5\" (UniqueName: \"kubernetes.io/projected/b9a6256a-adf8-4648-bc08-825f3d89a607-kube-api-access-9w7z5\") pod \"collect-profiles-29484750-cfdm7\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.507733 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:00 crc kubenswrapper[4874]: I0122 12:30:00.745525 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7"] Jan 22 12:30:01 crc kubenswrapper[4874]: I0122 12:30:01.156241 4874 generic.go:334] "Generic (PLEG): container finished" podID="b9a6256a-adf8-4648-bc08-825f3d89a607" containerID="d1dd2178451280b215a757384dff18acd9ee41c4da8e5236e7ad4dc659c2536a" exitCode=0 Jan 22 12:30:01 crc kubenswrapper[4874]: I0122 12:30:01.156361 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" event={"ID":"b9a6256a-adf8-4648-bc08-825f3d89a607","Type":"ContainerDied","Data":"d1dd2178451280b215a757384dff18acd9ee41c4da8e5236e7ad4dc659c2536a"} Jan 22 12:30:01 crc kubenswrapper[4874]: I0122 12:30:01.156592 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" event={"ID":"b9a6256a-adf8-4648-bc08-825f3d89a607","Type":"ContainerStarted","Data":"0edd124508f1e34abd9dac5ed4a94b67e8e7f638376944f4ed579e96dae37a7e"} Jan 22 12:30:02 crc kubenswrapper[4874]: I0122 12:30:02.499731 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:02 crc kubenswrapper[4874]: I0122 12:30:02.559908 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w7z5\" (UniqueName: \"kubernetes.io/projected/b9a6256a-adf8-4648-bc08-825f3d89a607-kube-api-access-9w7z5\") pod \"b9a6256a-adf8-4648-bc08-825f3d89a607\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " Jan 22 12:30:02 crc kubenswrapper[4874]: I0122 12:30:02.559987 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9a6256a-adf8-4648-bc08-825f3d89a607-secret-volume\") pod \"b9a6256a-adf8-4648-bc08-825f3d89a607\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " Jan 22 12:30:02 crc kubenswrapper[4874]: I0122 12:30:02.560029 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9a6256a-adf8-4648-bc08-825f3d89a607-config-volume\") pod \"b9a6256a-adf8-4648-bc08-825f3d89a607\" (UID: \"b9a6256a-adf8-4648-bc08-825f3d89a607\") " Jan 22 12:30:02 crc kubenswrapper[4874]: I0122 12:30:02.560697 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9a6256a-adf8-4648-bc08-825f3d89a607-config-volume" (OuterVolumeSpecName: "config-volume") pod "b9a6256a-adf8-4648-bc08-825f3d89a607" (UID: "b9a6256a-adf8-4648-bc08-825f3d89a607"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:30:02 crc kubenswrapper[4874]: I0122 12:30:02.565569 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a6256a-adf8-4648-bc08-825f3d89a607-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b9a6256a-adf8-4648-bc08-825f3d89a607" (UID: "b9a6256a-adf8-4648-bc08-825f3d89a607"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:30:02 crc kubenswrapper[4874]: I0122 12:30:02.565997 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a6256a-adf8-4648-bc08-825f3d89a607-kube-api-access-9w7z5" (OuterVolumeSpecName: "kube-api-access-9w7z5") pod "b9a6256a-adf8-4648-bc08-825f3d89a607" (UID: "b9a6256a-adf8-4648-bc08-825f3d89a607"). InnerVolumeSpecName "kube-api-access-9w7z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:30:02 crc kubenswrapper[4874]: I0122 12:30:02.661622 4874 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9a6256a-adf8-4648-bc08-825f3d89a607-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 12:30:02 crc kubenswrapper[4874]: I0122 12:30:02.661667 4874 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9a6256a-adf8-4648-bc08-825f3d89a607-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 12:30:02 crc kubenswrapper[4874]: I0122 12:30:02.661687 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w7z5\" (UniqueName: \"kubernetes.io/projected/b9a6256a-adf8-4648-bc08-825f3d89a607-kube-api-access-9w7z5\") on node \"crc\" DevicePath \"\"" Jan 22 12:30:03 crc kubenswrapper[4874]: I0122 12:30:03.176926 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" event={"ID":"b9a6256a-adf8-4648-bc08-825f3d89a607","Type":"ContainerDied","Data":"0edd124508f1e34abd9dac5ed4a94b67e8e7f638376944f4ed579e96dae37a7e"} Jan 22 12:30:03 crc kubenswrapper[4874]: I0122 12:30:03.177717 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0edd124508f1e34abd9dac5ed4a94b67e8e7f638376944f4ed579e96dae37a7e" Jan 22 12:30:03 crc kubenswrapper[4874]: I0122 12:30:03.177059 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484750-cfdm7" Jan 22 12:30:03 crc kubenswrapper[4874]: I0122 12:30:03.579154 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn"] Jan 22 12:30:03 crc kubenswrapper[4874]: I0122 12:30:03.590583 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484705-xlhwn"] Jan 22 12:30:04 crc kubenswrapper[4874]: I0122 12:30:04.725755 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce45c6ff-601e-4b12-97b6-4737304db2d7" path="/var/lib/kubelet/pods/ce45c6ff-601e-4b12-97b6-4737304db2d7/volumes" Jan 22 12:30:09 crc kubenswrapper[4874]: I0122 12:30:09.715836 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:30:09 crc kubenswrapper[4874]: E0122 12:30:09.716480 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:30:23 crc kubenswrapper[4874]: I0122 12:30:23.716177 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:30:23 crc kubenswrapper[4874]: E0122 12:30:23.717331 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:30:33 crc kubenswrapper[4874]: I0122 12:30:33.724969 4874 scope.go:117] "RemoveContainer" containerID="e4f9ca4f65548ceb4a0c2ab28cc0dcb70d2bfed6abf0e18366dac77436253610" Jan 22 12:30:38 crc kubenswrapper[4874]: I0122 12:30:38.717226 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:30:38 crc kubenswrapper[4874]: E0122 12:30:38.718073 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:30:51 crc kubenswrapper[4874]: I0122 12:30:51.716127 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:30:51 crc kubenswrapper[4874]: E0122 12:30:51.716751 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:31:04 crc kubenswrapper[4874]: I0122 12:31:04.715762 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:31:04 crc kubenswrapper[4874]: E0122 12:31:04.716686 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:31:17 crc kubenswrapper[4874]: I0122 12:31:17.716204 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:31:17 crc kubenswrapper[4874]: E0122 12:31:17.716992 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:31:31 crc kubenswrapper[4874]: I0122 12:31:31.716855 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:31:31 crc kubenswrapper[4874]: E0122 12:31:31.717824 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:31:42 crc kubenswrapper[4874]: I0122 12:31:42.715886 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:31:42 crc kubenswrapper[4874]: E0122 12:31:42.716701 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:31:53 crc kubenswrapper[4874]: I0122 12:31:53.718363 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:31:53 crc kubenswrapper[4874]: E0122 12:31:53.719228 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:32:06 crc kubenswrapper[4874]: I0122 12:32:06.731929 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:32:06 crc kubenswrapper[4874]: E0122 12:32:06.733439 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:32:17 crc kubenswrapper[4874]: I0122 12:32:17.717961 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:32:17 crc kubenswrapper[4874]: E0122 12:32:17.720152 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:32:32 crc kubenswrapper[4874]: I0122 12:32:32.718714 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:32:32 crc kubenswrapper[4874]: E0122 12:32:32.721049 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:32:45 crc kubenswrapper[4874]: I0122 12:32:45.717021 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:32:45 crc kubenswrapper[4874]: E0122 12:32:45.717861 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:32:57 crc kubenswrapper[4874]: I0122 12:32:57.715999 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:32:57 crc kubenswrapper[4874]: E0122 12:32:57.717072 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:33:08 crc kubenswrapper[4874]: I0122 12:33:08.723968 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:33:08 crc kubenswrapper[4874]: E0122 12:33:08.725163 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:33:19 crc kubenswrapper[4874]: I0122 12:33:19.716363 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:33:19 crc kubenswrapper[4874]: E0122 12:33:19.717325 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:33:31 crc kubenswrapper[4874]: I0122 12:33:31.715692 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:33:31 crc kubenswrapper[4874]: E0122 12:33:31.716636 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.020914 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dxtkp"] Jan 22 12:33:41 crc kubenswrapper[4874]: E0122 12:33:41.021790 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a6256a-adf8-4648-bc08-825f3d89a607" containerName="collect-profiles" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.021808 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a6256a-adf8-4648-bc08-825f3d89a607" containerName="collect-profiles" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.021993 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a6256a-adf8-4648-bc08-825f3d89a607" containerName="collect-profiles" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.029452 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.046091 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dxtkp"] Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.136525 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-utilities\") pod \"certified-operators-dxtkp\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.136598 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-catalog-content\") pod \"certified-operators-dxtkp\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.136630 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv8s9\" (UniqueName: \"kubernetes.io/projected/d4178041-14ad-4312-8200-0bc023576da8-kube-api-access-rv8s9\") pod \"certified-operators-dxtkp\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.238029 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-utilities\") pod \"certified-operators-dxtkp\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.238098 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-catalog-content\") pod \"certified-operators-dxtkp\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.238123 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv8s9\" (UniqueName: \"kubernetes.io/projected/d4178041-14ad-4312-8200-0bc023576da8-kube-api-access-rv8s9\") pod \"certified-operators-dxtkp\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.238672 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-utilities\") pod \"certified-operators-dxtkp\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.238749 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-catalog-content\") pod \"certified-operators-dxtkp\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.267133 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv8s9\" (UniqueName: \"kubernetes.io/projected/d4178041-14ad-4312-8200-0bc023576da8-kube-api-access-rv8s9\") pod \"certified-operators-dxtkp\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.353314 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:41 crc kubenswrapper[4874]: I0122 12:33:41.627567 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dxtkp"] Jan 22 12:33:42 crc kubenswrapper[4874]: I0122 12:33:42.180250 4874 generic.go:334] "Generic (PLEG): container finished" podID="d4178041-14ad-4312-8200-0bc023576da8" containerID="03fa9881f901b676e4d5bfa7b35603a03b5daba1c553dfcb37bd784ddd5ae75a" exitCode=0 Jan 22 12:33:42 crc kubenswrapper[4874]: I0122 12:33:42.180355 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dxtkp" event={"ID":"d4178041-14ad-4312-8200-0bc023576da8","Type":"ContainerDied","Data":"03fa9881f901b676e4d5bfa7b35603a03b5daba1c553dfcb37bd784ddd5ae75a"} Jan 22 12:33:42 crc kubenswrapper[4874]: I0122 12:33:42.180539 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dxtkp" event={"ID":"d4178041-14ad-4312-8200-0bc023576da8","Type":"ContainerStarted","Data":"06511af0e67adbc96e0266bc1edddc83525d3b3fd4f5e3c4d71165cd927480e7"} Jan 22 12:33:42 crc kubenswrapper[4874]: I0122 12:33:42.182090 4874 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 12:33:43 crc kubenswrapper[4874]: I0122 12:33:43.188454 4874 generic.go:334] "Generic (PLEG): container finished" podID="d4178041-14ad-4312-8200-0bc023576da8" containerID="dbdc0cc6b40c096de74bde2acc2ca457b7210736c0d92071699dd82d24092d21" exitCode=0 Jan 22 12:33:43 crc kubenswrapper[4874]: I0122 12:33:43.188501 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dxtkp" event={"ID":"d4178041-14ad-4312-8200-0bc023576da8","Type":"ContainerDied","Data":"dbdc0cc6b40c096de74bde2acc2ca457b7210736c0d92071699dd82d24092d21"} Jan 22 12:33:44 crc kubenswrapper[4874]: I0122 12:33:44.197504 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dxtkp" event={"ID":"d4178041-14ad-4312-8200-0bc023576da8","Type":"ContainerStarted","Data":"13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae"} Jan 22 12:33:44 crc kubenswrapper[4874]: I0122 12:33:44.218849 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dxtkp" podStartSLOduration=2.843595423 podStartE2EDuration="4.218833562s" podCreationTimestamp="2026-01-22 12:33:40 +0000 UTC" firstStartedPulling="2026-01-22 12:33:42.181896454 +0000 UTC m=+3196.026967524" lastFinishedPulling="2026-01-22 12:33:43.557134583 +0000 UTC m=+3197.402205663" observedRunningTime="2026-01-22 12:33:44.212496516 +0000 UTC m=+3198.057567586" watchObservedRunningTime="2026-01-22 12:33:44.218833562 +0000 UTC m=+3198.063904622" Jan 22 12:33:46 crc kubenswrapper[4874]: I0122 12:33:46.721196 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:33:46 crc kubenswrapper[4874]: E0122 12:33:46.722580 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:33:51 crc kubenswrapper[4874]: I0122 12:33:51.354460 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:51 crc kubenswrapper[4874]: I0122 12:33:51.354998 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:51 crc kubenswrapper[4874]: I0122 12:33:51.412630 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:52 crc kubenswrapper[4874]: I0122 12:33:52.321058 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:52 crc kubenswrapper[4874]: I0122 12:33:52.383379 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dxtkp"] Jan 22 12:33:54 crc kubenswrapper[4874]: I0122 12:33:54.274159 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dxtkp" podUID="d4178041-14ad-4312-8200-0bc023576da8" containerName="registry-server" containerID="cri-o://13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae" gracePeriod=2 Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.255505 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.301970 4874 generic.go:334] "Generic (PLEG): container finished" podID="d4178041-14ad-4312-8200-0bc023576da8" containerID="13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae" exitCode=0 Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.302151 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dxtkp" event={"ID":"d4178041-14ad-4312-8200-0bc023576da8","Type":"ContainerDied","Data":"13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae"} Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.302708 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dxtkp" event={"ID":"d4178041-14ad-4312-8200-0bc023576da8","Type":"ContainerDied","Data":"06511af0e67adbc96e0266bc1edddc83525d3b3fd4f5e3c4d71165cd927480e7"} Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.302265 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dxtkp" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.302805 4874 scope.go:117] "RemoveContainer" containerID="13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.324047 4874 scope.go:117] "RemoveContainer" containerID="dbdc0cc6b40c096de74bde2acc2ca457b7210736c0d92071699dd82d24092d21" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.346971 4874 scope.go:117] "RemoveContainer" containerID="03fa9881f901b676e4d5bfa7b35603a03b5daba1c553dfcb37bd784ddd5ae75a" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.354627 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv8s9\" (UniqueName: \"kubernetes.io/projected/d4178041-14ad-4312-8200-0bc023576da8-kube-api-access-rv8s9\") pod \"d4178041-14ad-4312-8200-0bc023576da8\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.354748 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-catalog-content\") pod \"d4178041-14ad-4312-8200-0bc023576da8\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.354825 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-utilities\") pod \"d4178041-14ad-4312-8200-0bc023576da8\" (UID: \"d4178041-14ad-4312-8200-0bc023576da8\") " Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.357690 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-utilities" (OuterVolumeSpecName: "utilities") pod "d4178041-14ad-4312-8200-0bc023576da8" (UID: "d4178041-14ad-4312-8200-0bc023576da8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.362644 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4178041-14ad-4312-8200-0bc023576da8-kube-api-access-rv8s9" (OuterVolumeSpecName: "kube-api-access-rv8s9") pod "d4178041-14ad-4312-8200-0bc023576da8" (UID: "d4178041-14ad-4312-8200-0bc023576da8"). InnerVolumeSpecName "kube-api-access-rv8s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.369387 4874 scope.go:117] "RemoveContainer" containerID="13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae" Jan 22 12:33:55 crc kubenswrapper[4874]: E0122 12:33:55.370140 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae\": container with ID starting with 13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae not found: ID does not exist" containerID="13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.370319 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae"} err="failed to get container status \"13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae\": rpc error: code = NotFound desc = could not find container \"13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae\": container with ID starting with 13573dc8f3c42b06e2e33121c0eb1d276a657f1c2afdd0ab2441688bb3844eae not found: ID does not exist" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.370562 4874 scope.go:117] "RemoveContainer" containerID="dbdc0cc6b40c096de74bde2acc2ca457b7210736c0d92071699dd82d24092d21" Jan 22 12:33:55 crc kubenswrapper[4874]: E0122 12:33:55.371110 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbdc0cc6b40c096de74bde2acc2ca457b7210736c0d92071699dd82d24092d21\": container with ID starting with dbdc0cc6b40c096de74bde2acc2ca457b7210736c0d92071699dd82d24092d21 not found: ID does not exist" containerID="dbdc0cc6b40c096de74bde2acc2ca457b7210736c0d92071699dd82d24092d21" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.371335 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbdc0cc6b40c096de74bde2acc2ca457b7210736c0d92071699dd82d24092d21"} err="failed to get container status \"dbdc0cc6b40c096de74bde2acc2ca457b7210736c0d92071699dd82d24092d21\": rpc error: code = NotFound desc = could not find container \"dbdc0cc6b40c096de74bde2acc2ca457b7210736c0d92071699dd82d24092d21\": container with ID starting with dbdc0cc6b40c096de74bde2acc2ca457b7210736c0d92071699dd82d24092d21 not found: ID does not exist" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.371530 4874 scope.go:117] "RemoveContainer" containerID="03fa9881f901b676e4d5bfa7b35603a03b5daba1c553dfcb37bd784ddd5ae75a" Jan 22 12:33:55 crc kubenswrapper[4874]: E0122 12:33:55.372095 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03fa9881f901b676e4d5bfa7b35603a03b5daba1c553dfcb37bd784ddd5ae75a\": container with ID starting with 03fa9881f901b676e4d5bfa7b35603a03b5daba1c553dfcb37bd784ddd5ae75a not found: ID does not exist" containerID="03fa9881f901b676e4d5bfa7b35603a03b5daba1c553dfcb37bd784ddd5ae75a" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.372284 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03fa9881f901b676e4d5bfa7b35603a03b5daba1c553dfcb37bd784ddd5ae75a"} err="failed to get container status \"03fa9881f901b676e4d5bfa7b35603a03b5daba1c553dfcb37bd784ddd5ae75a\": rpc error: code = NotFound desc = could not find container \"03fa9881f901b676e4d5bfa7b35603a03b5daba1c553dfcb37bd784ddd5ae75a\": container with ID starting with 03fa9881f901b676e4d5bfa7b35603a03b5daba1c553dfcb37bd784ddd5ae75a not found: ID does not exist" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.408454 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4178041-14ad-4312-8200-0bc023576da8" (UID: "d4178041-14ad-4312-8200-0bc023576da8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.456118 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv8s9\" (UniqueName: \"kubernetes.io/projected/d4178041-14ad-4312-8200-0bc023576da8-kube-api-access-rv8s9\") on node \"crc\" DevicePath \"\"" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.456149 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.456159 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4178041-14ad-4312-8200-0bc023576da8-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.646902 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dxtkp"] Jan 22 12:33:55 crc kubenswrapper[4874]: I0122 12:33:55.660348 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dxtkp"] Jan 22 12:33:56 crc kubenswrapper[4874]: I0122 12:33:56.730355 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4178041-14ad-4312-8200-0bc023576da8" path="/var/lib/kubelet/pods/d4178041-14ad-4312-8200-0bc023576da8/volumes" Jan 22 12:33:57 crc kubenswrapper[4874]: I0122 12:33:57.717774 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:33:57 crc kubenswrapper[4874]: E0122 12:33:57.718183 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:34:11 crc kubenswrapper[4874]: I0122 12:34:11.716601 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:34:11 crc kubenswrapper[4874]: E0122 12:34:11.719881 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:34:24 crc kubenswrapper[4874]: I0122 12:34:24.716676 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:34:24 crc kubenswrapper[4874]: E0122 12:34:24.717538 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:34:38 crc kubenswrapper[4874]: I0122 12:34:38.716123 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:34:38 crc kubenswrapper[4874]: E0122 12:34:38.716907 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:34:49 crc kubenswrapper[4874]: I0122 12:34:49.716390 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:34:50 crc kubenswrapper[4874]: I0122 12:34:50.861989 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"07181174f3dd294d8a33813c347d0b5e89eda4823fa5474d3fdedb3ef1a9e13a"} Jan 22 12:37:13 crc kubenswrapper[4874]: I0122 12:37:13.521010 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:37:13 crc kubenswrapper[4874]: I0122 12:37:13.521602 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:37:37 crc kubenswrapper[4874]: I0122 12:37:37.998536 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bdp5n"] Jan 22 12:37:38 crc kubenswrapper[4874]: E0122 12:37:37.999351 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4178041-14ad-4312-8200-0bc023576da8" containerName="extract-content" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:37.999367 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4178041-14ad-4312-8200-0bc023576da8" containerName="extract-content" Jan 22 12:37:38 crc kubenswrapper[4874]: E0122 12:37:37.999412 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4178041-14ad-4312-8200-0bc023576da8" containerName="extract-utilities" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:37.999423 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4178041-14ad-4312-8200-0bc023576da8" containerName="extract-utilities" Jan 22 12:37:38 crc kubenswrapper[4874]: E0122 12:37:37.999458 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4178041-14ad-4312-8200-0bc023576da8" containerName="registry-server" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:37.999465 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4178041-14ad-4312-8200-0bc023576da8" containerName="registry-server" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:37.999628 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4178041-14ad-4312-8200-0bc023576da8" containerName="registry-server" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.000697 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.017903 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bdp5n"] Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.148794 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5p5c\" (UniqueName: \"kubernetes.io/projected/1fb7aed4-6e72-4661-bc08-fad9bf646202-kube-api-access-p5p5c\") pod \"community-operators-bdp5n\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.148972 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-utilities\") pod \"community-operators-bdp5n\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.149000 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-catalog-content\") pod \"community-operators-bdp5n\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.250806 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5p5c\" (UniqueName: \"kubernetes.io/projected/1fb7aed4-6e72-4661-bc08-fad9bf646202-kube-api-access-p5p5c\") pod \"community-operators-bdp5n\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.251232 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-utilities\") pod \"community-operators-bdp5n\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.251256 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-catalog-content\") pod \"community-operators-bdp5n\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.251684 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-utilities\") pod \"community-operators-bdp5n\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.251775 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-catalog-content\") pod \"community-operators-bdp5n\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.269979 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5p5c\" (UniqueName: \"kubernetes.io/projected/1fb7aed4-6e72-4661-bc08-fad9bf646202-kube-api-access-p5p5c\") pod \"community-operators-bdp5n\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.318143 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:38 crc kubenswrapper[4874]: I0122 12:37:38.607530 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bdp5n"] Jan 22 12:37:39 crc kubenswrapper[4874]: I0122 12:37:39.398344 4874 generic.go:334] "Generic (PLEG): container finished" podID="1fb7aed4-6e72-4661-bc08-fad9bf646202" containerID="a3fb8abd9747c4bb400b1d4533e2a33a5f1808df7be49594ff6e9e5db4c9aa85" exitCode=0 Jan 22 12:37:39 crc kubenswrapper[4874]: I0122 12:37:39.398421 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdp5n" event={"ID":"1fb7aed4-6e72-4661-bc08-fad9bf646202","Type":"ContainerDied","Data":"a3fb8abd9747c4bb400b1d4533e2a33a5f1808df7be49594ff6e9e5db4c9aa85"} Jan 22 12:37:39 crc kubenswrapper[4874]: I0122 12:37:39.398633 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdp5n" event={"ID":"1fb7aed4-6e72-4661-bc08-fad9bf646202","Type":"ContainerStarted","Data":"d9a63872957764bb9bbe2995e7ec36fa28bf1580fa162fc26a40162b39a4f27e"} Jan 22 12:37:41 crc kubenswrapper[4874]: I0122 12:37:41.419568 4874 generic.go:334] "Generic (PLEG): container finished" podID="1fb7aed4-6e72-4661-bc08-fad9bf646202" containerID="0983b99ba4d62c33ac972b4d48dce835b2a92adc61a70384ce59116f8ff747e8" exitCode=0 Jan 22 12:37:41 crc kubenswrapper[4874]: I0122 12:37:41.419984 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdp5n" event={"ID":"1fb7aed4-6e72-4661-bc08-fad9bf646202","Type":"ContainerDied","Data":"0983b99ba4d62c33ac972b4d48dce835b2a92adc61a70384ce59116f8ff747e8"} Jan 22 12:37:42 crc kubenswrapper[4874]: I0122 12:37:42.430601 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdp5n" event={"ID":"1fb7aed4-6e72-4661-bc08-fad9bf646202","Type":"ContainerStarted","Data":"e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e"} Jan 22 12:37:42 crc kubenswrapper[4874]: I0122 12:37:42.461048 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bdp5n" podStartSLOduration=2.92455063 podStartE2EDuration="5.461027197s" podCreationTimestamp="2026-01-22 12:37:37 +0000 UTC" firstStartedPulling="2026-01-22 12:37:39.400428431 +0000 UTC m=+3433.245499501" lastFinishedPulling="2026-01-22 12:37:41.936904978 +0000 UTC m=+3435.781976068" observedRunningTime="2026-01-22 12:37:42.458699216 +0000 UTC m=+3436.303770296" watchObservedRunningTime="2026-01-22 12:37:42.461027197 +0000 UTC m=+3436.306098267" Jan 22 12:37:43 crc kubenswrapper[4874]: I0122 12:37:43.520449 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:37:43 crc kubenswrapper[4874]: I0122 12:37:43.520723 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:37:48 crc kubenswrapper[4874]: I0122 12:37:48.318686 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:48 crc kubenswrapper[4874]: I0122 12:37:48.319885 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:48 crc kubenswrapper[4874]: I0122 12:37:48.379541 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:48 crc kubenswrapper[4874]: I0122 12:37:48.517066 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:48 crc kubenswrapper[4874]: I0122 12:37:48.617867 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bdp5n"] Jan 22 12:37:50 crc kubenswrapper[4874]: I0122 12:37:50.486538 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bdp5n" podUID="1fb7aed4-6e72-4661-bc08-fad9bf646202" containerName="registry-server" containerID="cri-o://e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e" gracePeriod=2 Jan 22 12:37:51 crc kubenswrapper[4874]: I0122 12:37:51.942068 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.065884 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5p5c\" (UniqueName: \"kubernetes.io/projected/1fb7aed4-6e72-4661-bc08-fad9bf646202-kube-api-access-p5p5c\") pod \"1fb7aed4-6e72-4661-bc08-fad9bf646202\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.066053 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-catalog-content\") pod \"1fb7aed4-6e72-4661-bc08-fad9bf646202\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.066119 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-utilities\") pod \"1fb7aed4-6e72-4661-bc08-fad9bf646202\" (UID: \"1fb7aed4-6e72-4661-bc08-fad9bf646202\") " Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.067193 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-utilities" (OuterVolumeSpecName: "utilities") pod "1fb7aed4-6e72-4661-bc08-fad9bf646202" (UID: "1fb7aed4-6e72-4661-bc08-fad9bf646202"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.070945 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb7aed4-6e72-4661-bc08-fad9bf646202-kube-api-access-p5p5c" (OuterVolumeSpecName: "kube-api-access-p5p5c") pod "1fb7aed4-6e72-4661-bc08-fad9bf646202" (UID: "1fb7aed4-6e72-4661-bc08-fad9bf646202"). InnerVolumeSpecName "kube-api-access-p5p5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.121559 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1fb7aed4-6e72-4661-bc08-fad9bf646202" (UID: "1fb7aed4-6e72-4661-bc08-fad9bf646202"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.168639 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5p5c\" (UniqueName: \"kubernetes.io/projected/1fb7aed4-6e72-4661-bc08-fad9bf646202-kube-api-access-p5p5c\") on node \"crc\" DevicePath \"\"" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.168782 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.168813 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb7aed4-6e72-4661-bc08-fad9bf646202-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.510196 4874 generic.go:334] "Generic (PLEG): container finished" podID="1fb7aed4-6e72-4661-bc08-fad9bf646202" containerID="e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e" exitCode=0 Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.510243 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdp5n" event={"ID":"1fb7aed4-6e72-4661-bc08-fad9bf646202","Type":"ContainerDied","Data":"e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e"} Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.510280 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdp5n" event={"ID":"1fb7aed4-6e72-4661-bc08-fad9bf646202","Type":"ContainerDied","Data":"d9a63872957764bb9bbe2995e7ec36fa28bf1580fa162fc26a40162b39a4f27e"} Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.510302 4874 scope.go:117] "RemoveContainer" containerID="e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.510319 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdp5n" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.530078 4874 scope.go:117] "RemoveContainer" containerID="0983b99ba4d62c33ac972b4d48dce835b2a92adc61a70384ce59116f8ff747e8" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.554636 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bdp5n"] Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.569086 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bdp5n"] Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.571472 4874 scope.go:117] "RemoveContainer" containerID="a3fb8abd9747c4bb400b1d4533e2a33a5f1808df7be49594ff6e9e5db4c9aa85" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.592370 4874 scope.go:117] "RemoveContainer" containerID="e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e" Jan 22 12:37:52 crc kubenswrapper[4874]: E0122 12:37:52.592820 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e\": container with ID starting with e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e not found: ID does not exist" containerID="e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.592849 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e"} err="failed to get container status \"e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e\": rpc error: code = NotFound desc = could not find container \"e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e\": container with ID starting with e440d1b5a98a70fb0140d96828eeff03c8761c2df34a5714a0f47a188c26f27e not found: ID does not exist" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.592869 4874 scope.go:117] "RemoveContainer" containerID="0983b99ba4d62c33ac972b4d48dce835b2a92adc61a70384ce59116f8ff747e8" Jan 22 12:37:52 crc kubenswrapper[4874]: E0122 12:37:52.593150 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0983b99ba4d62c33ac972b4d48dce835b2a92adc61a70384ce59116f8ff747e8\": container with ID starting with 0983b99ba4d62c33ac972b4d48dce835b2a92adc61a70384ce59116f8ff747e8 not found: ID does not exist" containerID="0983b99ba4d62c33ac972b4d48dce835b2a92adc61a70384ce59116f8ff747e8" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.593179 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0983b99ba4d62c33ac972b4d48dce835b2a92adc61a70384ce59116f8ff747e8"} err="failed to get container status \"0983b99ba4d62c33ac972b4d48dce835b2a92adc61a70384ce59116f8ff747e8\": rpc error: code = NotFound desc = could not find container \"0983b99ba4d62c33ac972b4d48dce835b2a92adc61a70384ce59116f8ff747e8\": container with ID starting with 0983b99ba4d62c33ac972b4d48dce835b2a92adc61a70384ce59116f8ff747e8 not found: ID does not exist" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.593193 4874 scope.go:117] "RemoveContainer" containerID="a3fb8abd9747c4bb400b1d4533e2a33a5f1808df7be49594ff6e9e5db4c9aa85" Jan 22 12:37:52 crc kubenswrapper[4874]: E0122 12:37:52.593577 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3fb8abd9747c4bb400b1d4533e2a33a5f1808df7be49594ff6e9e5db4c9aa85\": container with ID starting with a3fb8abd9747c4bb400b1d4533e2a33a5f1808df7be49594ff6e9e5db4c9aa85 not found: ID does not exist" containerID="a3fb8abd9747c4bb400b1d4533e2a33a5f1808df7be49594ff6e9e5db4c9aa85" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.593679 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3fb8abd9747c4bb400b1d4533e2a33a5f1808df7be49594ff6e9e5db4c9aa85"} err="failed to get container status \"a3fb8abd9747c4bb400b1d4533e2a33a5f1808df7be49594ff6e9e5db4c9aa85\": rpc error: code = NotFound desc = could not find container \"a3fb8abd9747c4bb400b1d4533e2a33a5f1808df7be49594ff6e9e5db4c9aa85\": container with ID starting with a3fb8abd9747c4bb400b1d4533e2a33a5f1808df7be49594ff6e9e5db4c9aa85 not found: ID does not exist" Jan 22 12:37:52 crc kubenswrapper[4874]: I0122 12:37:52.724553 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fb7aed4-6e72-4661-bc08-fad9bf646202" path="/var/lib/kubelet/pods/1fb7aed4-6e72-4661-bc08-fad9bf646202/volumes" Jan 22 12:38:13 crc kubenswrapper[4874]: I0122 12:38:13.520444 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:38:13 crc kubenswrapper[4874]: I0122 12:38:13.520868 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:38:13 crc kubenswrapper[4874]: I0122 12:38:13.520907 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 12:38:13 crc kubenswrapper[4874]: I0122 12:38:13.521627 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07181174f3dd294d8a33813c347d0b5e89eda4823fa5474d3fdedb3ef1a9e13a"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 12:38:13 crc kubenswrapper[4874]: I0122 12:38:13.521670 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://07181174f3dd294d8a33813c347d0b5e89eda4823fa5474d3fdedb3ef1a9e13a" gracePeriod=600 Jan 22 12:38:13 crc kubenswrapper[4874]: I0122 12:38:13.705577 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="07181174f3dd294d8a33813c347d0b5e89eda4823fa5474d3fdedb3ef1a9e13a" exitCode=0 Jan 22 12:38:13 crc kubenswrapper[4874]: I0122 12:38:13.705632 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"07181174f3dd294d8a33813c347d0b5e89eda4823fa5474d3fdedb3ef1a9e13a"} Jan 22 12:38:13 crc kubenswrapper[4874]: I0122 12:38:13.705669 4874 scope.go:117] "RemoveContainer" containerID="01df6b3be23e0ceccc44bb801500d768b66322dd1258e6f93bd6e67bc3e1efd7" Jan 22 12:38:14 crc kubenswrapper[4874]: I0122 12:38:14.732144 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b"} Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.471169 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9n79x"] Jan 22 12:39:18 crc kubenswrapper[4874]: E0122 12:39:18.472380 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb7aed4-6e72-4661-bc08-fad9bf646202" containerName="registry-server" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.472433 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb7aed4-6e72-4661-bc08-fad9bf646202" containerName="registry-server" Jan 22 12:39:18 crc kubenswrapper[4874]: E0122 12:39:18.472456 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb7aed4-6e72-4661-bc08-fad9bf646202" containerName="extract-utilities" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.472468 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb7aed4-6e72-4661-bc08-fad9bf646202" containerName="extract-utilities" Jan 22 12:39:18 crc kubenswrapper[4874]: E0122 12:39:18.472488 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb7aed4-6e72-4661-bc08-fad9bf646202" containerName="extract-content" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.472500 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb7aed4-6e72-4661-bc08-fad9bf646202" containerName="extract-content" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.472750 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fb7aed4-6e72-4661-bc08-fad9bf646202" containerName="registry-server" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.474688 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.482888 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9n79x"] Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.582275 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-utilities\") pod \"redhat-operators-9n79x\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.582349 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7gmt\" (UniqueName: \"kubernetes.io/projected/b7b37623-67a8-4f21-9152-c51710b5a5e5-kube-api-access-w7gmt\") pod \"redhat-operators-9n79x\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.582656 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-catalog-content\") pod \"redhat-operators-9n79x\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.684234 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-utilities\") pod \"redhat-operators-9n79x\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.684328 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7gmt\" (UniqueName: \"kubernetes.io/projected/b7b37623-67a8-4f21-9152-c51710b5a5e5-kube-api-access-w7gmt\") pod \"redhat-operators-9n79x\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.684445 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-catalog-content\") pod \"redhat-operators-9n79x\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.685062 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-catalog-content\") pod \"redhat-operators-9n79x\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.685204 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-utilities\") pod \"redhat-operators-9n79x\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.702500 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7gmt\" (UniqueName: \"kubernetes.io/projected/b7b37623-67a8-4f21-9152-c51710b5a5e5-kube-api-access-w7gmt\") pod \"redhat-operators-9n79x\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:18 crc kubenswrapper[4874]: I0122 12:39:18.812313 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:19 crc kubenswrapper[4874]: I0122 12:39:19.022422 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9n79x"] Jan 22 12:39:19 crc kubenswrapper[4874]: I0122 12:39:19.260150 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n79x" event={"ID":"b7b37623-67a8-4f21-9152-c51710b5a5e5","Type":"ContainerStarted","Data":"9f023910baea6188cb4dcf599c845beb6ab5eed90edccfda962087fc408ed0b0"} Jan 22 12:39:20 crc kubenswrapper[4874]: I0122 12:39:20.266420 4874 generic.go:334] "Generic (PLEG): container finished" podID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerID="5f40bc937a97917b661dbc429c0bc3e2725cafe729004e75691d3832cb5d5cff" exitCode=0 Jan 22 12:39:20 crc kubenswrapper[4874]: I0122 12:39:20.266457 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n79x" event={"ID":"b7b37623-67a8-4f21-9152-c51710b5a5e5","Type":"ContainerDied","Data":"5f40bc937a97917b661dbc429c0bc3e2725cafe729004e75691d3832cb5d5cff"} Jan 22 12:39:20 crc kubenswrapper[4874]: I0122 12:39:20.268080 4874 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 12:39:22 crc kubenswrapper[4874]: I0122 12:39:22.282936 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n79x" event={"ID":"b7b37623-67a8-4f21-9152-c51710b5a5e5","Type":"ContainerStarted","Data":"a394968505723f72cc7683f448a2532ffda3dda4790de0130c838c94033d17ea"} Jan 22 12:39:23 crc kubenswrapper[4874]: I0122 12:39:23.291366 4874 generic.go:334] "Generic (PLEG): container finished" podID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerID="a394968505723f72cc7683f448a2532ffda3dda4790de0130c838c94033d17ea" exitCode=0 Jan 22 12:39:23 crc kubenswrapper[4874]: I0122 12:39:23.291420 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n79x" event={"ID":"b7b37623-67a8-4f21-9152-c51710b5a5e5","Type":"ContainerDied","Data":"a394968505723f72cc7683f448a2532ffda3dda4790de0130c838c94033d17ea"} Jan 22 12:39:26 crc kubenswrapper[4874]: I0122 12:39:26.319815 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n79x" event={"ID":"b7b37623-67a8-4f21-9152-c51710b5a5e5","Type":"ContainerStarted","Data":"4e959bd367b219aa447bff1d69bec8f685cc10afcc77b393c6a31ac1dcf1aa0d"} Jan 22 12:39:26 crc kubenswrapper[4874]: I0122 12:39:26.344167 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9n79x" podStartSLOduration=3.411750087 podStartE2EDuration="8.344148638s" podCreationTimestamp="2026-01-22 12:39:18 +0000 UTC" firstStartedPulling="2026-01-22 12:39:20.267839685 +0000 UTC m=+3534.112910755" lastFinishedPulling="2026-01-22 12:39:25.200238226 +0000 UTC m=+3539.045309306" observedRunningTime="2026-01-22 12:39:26.344063505 +0000 UTC m=+3540.189134575" watchObservedRunningTime="2026-01-22 12:39:26.344148638 +0000 UTC m=+3540.189219718" Jan 22 12:39:28 crc kubenswrapper[4874]: I0122 12:39:28.812628 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:28 crc kubenswrapper[4874]: I0122 12:39:28.813990 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:29 crc kubenswrapper[4874]: I0122 12:39:29.863202 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9n79x" podUID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerName="registry-server" probeResult="failure" output=< Jan 22 12:39:29 crc kubenswrapper[4874]: timeout: failed to connect service ":50051" within 1s Jan 22 12:39:29 crc kubenswrapper[4874]: > Jan 22 12:39:38 crc kubenswrapper[4874]: I0122 12:39:38.871920 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:38 crc kubenswrapper[4874]: I0122 12:39:38.951101 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.114580 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9n79x"] Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.115329 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9n79x" podUID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerName="registry-server" containerID="cri-o://4e959bd367b219aa447bff1d69bec8f685cc10afcc77b393c6a31ac1dcf1aa0d" gracePeriod=2 Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.457594 4874 generic.go:334] "Generic (PLEG): container finished" podID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerID="4e959bd367b219aa447bff1d69bec8f685cc10afcc77b393c6a31ac1dcf1aa0d" exitCode=0 Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.458038 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n79x" event={"ID":"b7b37623-67a8-4f21-9152-c51710b5a5e5","Type":"ContainerDied","Data":"4e959bd367b219aa447bff1d69bec8f685cc10afcc77b393c6a31ac1dcf1aa0d"} Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.619166 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.773029 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-catalog-content\") pod \"b7b37623-67a8-4f21-9152-c51710b5a5e5\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.773164 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7gmt\" (UniqueName: \"kubernetes.io/projected/b7b37623-67a8-4f21-9152-c51710b5a5e5-kube-api-access-w7gmt\") pod \"b7b37623-67a8-4f21-9152-c51710b5a5e5\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.773285 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-utilities\") pod \"b7b37623-67a8-4f21-9152-c51710b5a5e5\" (UID: \"b7b37623-67a8-4f21-9152-c51710b5a5e5\") " Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.775290 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-utilities" (OuterVolumeSpecName: "utilities") pod "b7b37623-67a8-4f21-9152-c51710b5a5e5" (UID: "b7b37623-67a8-4f21-9152-c51710b5a5e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.801709 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7b37623-67a8-4f21-9152-c51710b5a5e5-kube-api-access-w7gmt" (OuterVolumeSpecName: "kube-api-access-w7gmt") pod "b7b37623-67a8-4f21-9152-c51710b5a5e5" (UID: "b7b37623-67a8-4f21-9152-c51710b5a5e5"). InnerVolumeSpecName "kube-api-access-w7gmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.875802 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.875851 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7gmt\" (UniqueName: \"kubernetes.io/projected/b7b37623-67a8-4f21-9152-c51710b5a5e5-kube-api-access-w7gmt\") on node \"crc\" DevicePath \"\"" Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.961829 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7b37623-67a8-4f21-9152-c51710b5a5e5" (UID: "b7b37623-67a8-4f21-9152-c51710b5a5e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:39:41 crc kubenswrapper[4874]: I0122 12:39:41.978733 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b37623-67a8-4f21-9152-c51710b5a5e5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:39:42 crc kubenswrapper[4874]: I0122 12:39:42.466583 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n79x" event={"ID":"b7b37623-67a8-4f21-9152-c51710b5a5e5","Type":"ContainerDied","Data":"9f023910baea6188cb4dcf599c845beb6ab5eed90edccfda962087fc408ed0b0"} Jan 22 12:39:42 crc kubenswrapper[4874]: I0122 12:39:42.466921 4874 scope.go:117] "RemoveContainer" containerID="4e959bd367b219aa447bff1d69bec8f685cc10afcc77b393c6a31ac1dcf1aa0d" Jan 22 12:39:42 crc kubenswrapper[4874]: I0122 12:39:42.467118 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n79x" Jan 22 12:39:42 crc kubenswrapper[4874]: I0122 12:39:42.513632 4874 scope.go:117] "RemoveContainer" containerID="a394968505723f72cc7683f448a2532ffda3dda4790de0130c838c94033d17ea" Jan 22 12:39:42 crc kubenswrapper[4874]: I0122 12:39:42.530603 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9n79x"] Jan 22 12:39:42 crc kubenswrapper[4874]: I0122 12:39:42.538610 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9n79x"] Jan 22 12:39:42 crc kubenswrapper[4874]: I0122 12:39:42.543595 4874 scope.go:117] "RemoveContainer" containerID="5f40bc937a97917b661dbc429c0bc3e2725cafe729004e75691d3832cb5d5cff" Jan 22 12:39:42 crc kubenswrapper[4874]: I0122 12:39:42.738312 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7b37623-67a8-4f21-9152-c51710b5a5e5" path="/var/lib/kubelet/pods/b7b37623-67a8-4f21-9152-c51710b5a5e5/volumes" Jan 22 12:40:13 crc kubenswrapper[4874]: I0122 12:40:13.520960 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:40:13 crc kubenswrapper[4874]: I0122 12:40:13.521949 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:40:43 crc kubenswrapper[4874]: I0122 12:40:43.521139 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:40:43 crc kubenswrapper[4874]: I0122 12:40:43.521841 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:41:13 crc kubenswrapper[4874]: I0122 12:41:13.520854 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:41:13 crc kubenswrapper[4874]: I0122 12:41:13.523294 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:41:13 crc kubenswrapper[4874]: I0122 12:41:13.523882 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 12:41:13 crc kubenswrapper[4874]: I0122 12:41:13.530147 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 12:41:13 crc kubenswrapper[4874]: I0122 12:41:13.530689 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" gracePeriod=600 Jan 22 12:41:14 crc kubenswrapper[4874]: E0122 12:41:14.163658 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:41:14 crc kubenswrapper[4874]: I0122 12:41:14.227112 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" exitCode=0 Jan 22 12:41:14 crc kubenswrapper[4874]: I0122 12:41:14.227179 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b"} Jan 22 12:41:14 crc kubenswrapper[4874]: I0122 12:41:14.227230 4874 scope.go:117] "RemoveContainer" containerID="07181174f3dd294d8a33813c347d0b5e89eda4823fa5474d3fdedb3ef1a9e13a" Jan 22 12:41:14 crc kubenswrapper[4874]: I0122 12:41:14.229201 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:41:14 crc kubenswrapper[4874]: E0122 12:41:14.229737 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:41:26 crc kubenswrapper[4874]: I0122 12:41:26.762585 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:41:26 crc kubenswrapper[4874]: E0122 12:41:26.763615 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:41:37 crc kubenswrapper[4874]: I0122 12:41:37.715985 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:41:37 crc kubenswrapper[4874]: E0122 12:41:37.717028 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:41:49 crc kubenswrapper[4874]: I0122 12:41:49.716355 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:41:49 crc kubenswrapper[4874]: E0122 12:41:49.717171 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:42:00 crc kubenswrapper[4874]: I0122 12:42:00.717919 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:42:00 crc kubenswrapper[4874]: E0122 12:42:00.721012 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:42:14 crc kubenswrapper[4874]: I0122 12:42:14.716733 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:42:14 crc kubenswrapper[4874]: E0122 12:42:14.717713 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:42:25 crc kubenswrapper[4874]: I0122 12:42:25.716278 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:42:25 crc kubenswrapper[4874]: E0122 12:42:25.717247 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:42:40 crc kubenswrapper[4874]: I0122 12:42:40.717085 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:42:40 crc kubenswrapper[4874]: E0122 12:42:40.718490 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:42:53 crc kubenswrapper[4874]: I0122 12:42:53.717022 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:42:53 crc kubenswrapper[4874]: E0122 12:42:53.717748 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:43:04 crc kubenswrapper[4874]: I0122 12:43:04.716784 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:43:04 crc kubenswrapper[4874]: E0122 12:43:04.718005 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:43:19 crc kubenswrapper[4874]: I0122 12:43:19.716040 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:43:19 crc kubenswrapper[4874]: E0122 12:43:19.716889 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:43:30 crc kubenswrapper[4874]: I0122 12:43:30.719236 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:43:30 crc kubenswrapper[4874]: E0122 12:43:30.720764 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:43:44 crc kubenswrapper[4874]: I0122 12:43:44.716872 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:43:44 crc kubenswrapper[4874]: E0122 12:43:44.717914 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:43:58 crc kubenswrapper[4874]: I0122 12:43:58.716321 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:43:58 crc kubenswrapper[4874]: E0122 12:43:58.717269 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:44:13 crc kubenswrapper[4874]: I0122 12:44:13.716596 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:44:13 crc kubenswrapper[4874]: E0122 12:44:13.717257 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:44:24 crc kubenswrapper[4874]: I0122 12:44:24.716968 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:44:24 crc kubenswrapper[4874]: E0122 12:44:24.718152 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:44:38 crc kubenswrapper[4874]: I0122 12:44:38.720996 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:44:38 crc kubenswrapper[4874]: E0122 12:44:38.721677 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:44:48 crc kubenswrapper[4874]: I0122 12:44:48.936541 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7mj86"] Jan 22 12:44:48 crc kubenswrapper[4874]: E0122 12:44:48.937615 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerName="registry-server" Jan 22 12:44:48 crc kubenswrapper[4874]: I0122 12:44:48.937635 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerName="registry-server" Jan 22 12:44:48 crc kubenswrapper[4874]: E0122 12:44:48.937652 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerName="extract-content" Jan 22 12:44:48 crc kubenswrapper[4874]: I0122 12:44:48.937661 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerName="extract-content" Jan 22 12:44:48 crc kubenswrapper[4874]: E0122 12:44:48.937690 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerName="extract-utilities" Jan 22 12:44:48 crc kubenswrapper[4874]: I0122 12:44:48.937699 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerName="extract-utilities" Jan 22 12:44:48 crc kubenswrapper[4874]: I0122 12:44:48.937853 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7b37623-67a8-4f21-9152-c51710b5a5e5" containerName="registry-server" Jan 22 12:44:48 crc kubenswrapper[4874]: I0122 12:44:48.939126 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:48 crc kubenswrapper[4874]: I0122 12:44:48.944903 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mj86"] Jan 22 12:44:48 crc kubenswrapper[4874]: I0122 12:44:48.975179 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-catalog-content\") pod \"certified-operators-7mj86\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:48 crc kubenswrapper[4874]: I0122 12:44:48.975231 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-utilities\") pod \"certified-operators-7mj86\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:48 crc kubenswrapper[4874]: I0122 12:44:48.975418 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v87g\" (UniqueName: \"kubernetes.io/projected/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-kube-api-access-7v87g\") pod \"certified-operators-7mj86\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:49 crc kubenswrapper[4874]: I0122 12:44:49.076369 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-catalog-content\") pod \"certified-operators-7mj86\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:49 crc kubenswrapper[4874]: I0122 12:44:49.076432 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-utilities\") pod \"certified-operators-7mj86\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:49 crc kubenswrapper[4874]: I0122 12:44:49.076555 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v87g\" (UniqueName: \"kubernetes.io/projected/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-kube-api-access-7v87g\") pod \"certified-operators-7mj86\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:49 crc kubenswrapper[4874]: I0122 12:44:49.076914 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-utilities\") pod \"certified-operators-7mj86\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:49 crc kubenswrapper[4874]: I0122 12:44:49.077145 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-catalog-content\") pod \"certified-operators-7mj86\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:49 crc kubenswrapper[4874]: I0122 12:44:49.691557 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v87g\" (UniqueName: \"kubernetes.io/projected/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-kube-api-access-7v87g\") pod \"certified-operators-7mj86\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:49 crc kubenswrapper[4874]: I0122 12:44:49.868428 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:50 crc kubenswrapper[4874]: I0122 12:44:50.096561 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mj86"] Jan 22 12:44:50 crc kubenswrapper[4874]: I0122 12:44:50.196661 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mj86" event={"ID":"402518c1-cf9f-4e51-9e46-2c7d1527ce6d","Type":"ContainerStarted","Data":"8ae86083e447aa6ec881faf0e570a3a06ed57d31d32d7fb452cb02a25568d862"} Jan 22 12:44:50 crc kubenswrapper[4874]: I0122 12:44:50.717036 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:44:50 crc kubenswrapper[4874]: E0122 12:44:50.717815 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:44:51 crc kubenswrapper[4874]: I0122 12:44:51.208035 4874 generic.go:334] "Generic (PLEG): container finished" podID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" containerID="468a698adc17dc00dba716877ea0acfe03d366fef4507a99ac8e71a28acd0f19" exitCode=0 Jan 22 12:44:51 crc kubenswrapper[4874]: I0122 12:44:51.208100 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mj86" event={"ID":"402518c1-cf9f-4e51-9e46-2c7d1527ce6d","Type":"ContainerDied","Data":"468a698adc17dc00dba716877ea0acfe03d366fef4507a99ac8e71a28acd0f19"} Jan 22 12:44:51 crc kubenswrapper[4874]: I0122 12:44:51.211069 4874 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 12:44:52 crc kubenswrapper[4874]: I0122 12:44:52.218442 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mj86" event={"ID":"402518c1-cf9f-4e51-9e46-2c7d1527ce6d","Type":"ContainerStarted","Data":"ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8"} Jan 22 12:44:53 crc kubenswrapper[4874]: I0122 12:44:53.225003 4874 generic.go:334] "Generic (PLEG): container finished" podID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" containerID="ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8" exitCode=0 Jan 22 12:44:53 crc kubenswrapper[4874]: I0122 12:44:53.225055 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mj86" event={"ID":"402518c1-cf9f-4e51-9e46-2c7d1527ce6d","Type":"ContainerDied","Data":"ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8"} Jan 22 12:44:54 crc kubenswrapper[4874]: I0122 12:44:54.236690 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mj86" event={"ID":"402518c1-cf9f-4e51-9e46-2c7d1527ce6d","Type":"ContainerStarted","Data":"b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443"} Jan 22 12:44:54 crc kubenswrapper[4874]: I0122 12:44:54.267783 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7mj86" podStartSLOduration=3.835496477 podStartE2EDuration="6.267768058s" podCreationTimestamp="2026-01-22 12:44:48 +0000 UTC" firstStartedPulling="2026-01-22 12:44:51.210701639 +0000 UTC m=+3865.055772739" lastFinishedPulling="2026-01-22 12:44:53.64297325 +0000 UTC m=+3867.488044320" observedRunningTime="2026-01-22 12:44:54.263808045 +0000 UTC m=+3868.108879115" watchObservedRunningTime="2026-01-22 12:44:54.267768058 +0000 UTC m=+3868.112839128" Jan 22 12:44:59 crc kubenswrapper[4874]: I0122 12:44:59.869135 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:59 crc kubenswrapper[4874]: I0122 12:44:59.869876 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:44:59 crc kubenswrapper[4874]: I0122 12:44:59.919905 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.181628 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl"] Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.182386 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.185692 4874 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.185885 4874 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.201323 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl"] Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.232422 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fql5q\" (UniqueName: \"kubernetes.io/projected/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-kube-api-access-fql5q\") pod \"collect-profiles-29484765-kr6bl\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.232553 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-secret-volume\") pod \"collect-profiles-29484765-kr6bl\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.232607 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-config-volume\") pod \"collect-profiles-29484765-kr6bl\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.334553 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fql5q\" (UniqueName: \"kubernetes.io/projected/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-kube-api-access-fql5q\") pod \"collect-profiles-29484765-kr6bl\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.334851 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-secret-volume\") pod \"collect-profiles-29484765-kr6bl\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.335577 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-config-volume\") pod \"collect-profiles-29484765-kr6bl\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.336207 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-config-volume\") pod \"collect-profiles-29484765-kr6bl\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.336716 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.341413 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-secret-volume\") pod \"collect-profiles-29484765-kr6bl\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.365140 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fql5q\" (UniqueName: \"kubernetes.io/projected/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-kube-api-access-fql5q\") pod \"collect-profiles-29484765-kr6bl\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.391773 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7mj86"] Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.500031 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:00 crc kubenswrapper[4874]: I0122 12:45:00.701941 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl"] Jan 22 12:45:01 crc kubenswrapper[4874]: I0122 12:45:01.317953 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" event={"ID":"d729e7c6-4664-47cb-b4bd-9892e1cf5aec","Type":"ContainerStarted","Data":"d454170fcee92be58faf76410af524e6d12e151305a1d5537c188dd81c65eef4"} Jan 22 12:45:01 crc kubenswrapper[4874]: I0122 12:45:01.318103 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" event={"ID":"d729e7c6-4664-47cb-b4bd-9892e1cf5aec","Type":"ContainerStarted","Data":"909e7447c9e1e506cd07baba43a99e331ccc847b174d0ffcaffa649c2276d6a9"} Jan 22 12:45:01 crc kubenswrapper[4874]: I0122 12:45:01.339856 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" podStartSLOduration=1.3398280439999999 podStartE2EDuration="1.339828044s" podCreationTimestamp="2026-01-22 12:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 12:45:01.334784179 +0000 UTC m=+3875.179855289" watchObservedRunningTime="2026-01-22 12:45:01.339828044 +0000 UTC m=+3875.184899124" Jan 22 12:45:02 crc kubenswrapper[4874]: I0122 12:45:02.327781 4874 generic.go:334] "Generic (PLEG): container finished" podID="d729e7c6-4664-47cb-b4bd-9892e1cf5aec" containerID="d454170fcee92be58faf76410af524e6d12e151305a1d5537c188dd81c65eef4" exitCode=0 Jan 22 12:45:02 crc kubenswrapper[4874]: I0122 12:45:02.329478 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7mj86" podUID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" containerName="registry-server" containerID="cri-o://b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443" gracePeriod=2 Jan 22 12:45:02 crc kubenswrapper[4874]: I0122 12:45:02.328077 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" event={"ID":"d729e7c6-4664-47cb-b4bd-9892e1cf5aec","Type":"ContainerDied","Data":"d454170fcee92be58faf76410af524e6d12e151305a1d5537c188dd81c65eef4"} Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.159025 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.208267 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-utilities\") pod \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.208354 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-catalog-content\") pod \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.208388 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v87g\" (UniqueName: \"kubernetes.io/projected/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-kube-api-access-7v87g\") pod \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\" (UID: \"402518c1-cf9f-4e51-9e46-2c7d1527ce6d\") " Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.209252 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-utilities" (OuterVolumeSpecName: "utilities") pod "402518c1-cf9f-4e51-9e46-2c7d1527ce6d" (UID: "402518c1-cf9f-4e51-9e46-2c7d1527ce6d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.222315 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-kube-api-access-7v87g" (OuterVolumeSpecName: "kube-api-access-7v87g") pod "402518c1-cf9f-4e51-9e46-2c7d1527ce6d" (UID: "402518c1-cf9f-4e51-9e46-2c7d1527ce6d"). InnerVolumeSpecName "kube-api-access-7v87g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.269494 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "402518c1-cf9f-4e51-9e46-2c7d1527ce6d" (UID: "402518c1-cf9f-4e51-9e46-2c7d1527ce6d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.310024 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.310088 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.310104 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7v87g\" (UniqueName: \"kubernetes.io/projected/402518c1-cf9f-4e51-9e46-2c7d1527ce6d-kube-api-access-7v87g\") on node \"crc\" DevicePath \"\"" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.342256 4874 generic.go:334] "Generic (PLEG): container finished" podID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" containerID="b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443" exitCode=0 Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.342379 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mj86" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.342471 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mj86" event={"ID":"402518c1-cf9f-4e51-9e46-2c7d1527ce6d","Type":"ContainerDied","Data":"b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443"} Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.342541 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mj86" event={"ID":"402518c1-cf9f-4e51-9e46-2c7d1527ce6d","Type":"ContainerDied","Data":"8ae86083e447aa6ec881faf0e570a3a06ed57d31d32d7fb452cb02a25568d862"} Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.342564 4874 scope.go:117] "RemoveContainer" containerID="b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.367648 4874 scope.go:117] "RemoveContainer" containerID="ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.387268 4874 scope.go:117] "RemoveContainer" containerID="468a698adc17dc00dba716877ea0acfe03d366fef4507a99ac8e71a28acd0f19" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.404193 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7mj86"] Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.408559 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7mj86"] Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.522953 4874 scope.go:117] "RemoveContainer" containerID="b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443" Jan 22 12:45:03 crc kubenswrapper[4874]: E0122 12:45:03.523567 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443\": container with ID starting with b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443 not found: ID does not exist" containerID="b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.523649 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443"} err="failed to get container status \"b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443\": rpc error: code = NotFound desc = could not find container \"b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443\": container with ID starting with b79e90f22636cd0aed9aa68ea9ce648ca46b137121b7b51414479a1ad8658443 not found: ID does not exist" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.523701 4874 scope.go:117] "RemoveContainer" containerID="ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8" Jan 22 12:45:03 crc kubenswrapper[4874]: E0122 12:45:03.524128 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8\": container with ID starting with ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8 not found: ID does not exist" containerID="ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.524183 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8"} err="failed to get container status \"ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8\": rpc error: code = NotFound desc = could not find container \"ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8\": container with ID starting with ae2787145bd140f284b5eb542ae4e0ed93975463dadefb31e27e4813c69e93b8 not found: ID does not exist" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.524209 4874 scope.go:117] "RemoveContainer" containerID="468a698adc17dc00dba716877ea0acfe03d366fef4507a99ac8e71a28acd0f19" Jan 22 12:45:03 crc kubenswrapper[4874]: E0122 12:45:03.524504 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"468a698adc17dc00dba716877ea0acfe03d366fef4507a99ac8e71a28acd0f19\": container with ID starting with 468a698adc17dc00dba716877ea0acfe03d366fef4507a99ac8e71a28acd0f19 not found: ID does not exist" containerID="468a698adc17dc00dba716877ea0acfe03d366fef4507a99ac8e71a28acd0f19" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.524536 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"468a698adc17dc00dba716877ea0acfe03d366fef4507a99ac8e71a28acd0f19"} err="failed to get container status \"468a698adc17dc00dba716877ea0acfe03d366fef4507a99ac8e71a28acd0f19\": rpc error: code = NotFound desc = could not find container \"468a698adc17dc00dba716877ea0acfe03d366fef4507a99ac8e71a28acd0f19\": container with ID starting with 468a698adc17dc00dba716877ea0acfe03d366fef4507a99ac8e71a28acd0f19 not found: ID does not exist" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.681308 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.716547 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:45:03 crc kubenswrapper[4874]: E0122 12:45:03.716861 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.833911 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fql5q\" (UniqueName: \"kubernetes.io/projected/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-kube-api-access-fql5q\") pod \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.834041 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-config-volume\") pod \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.834116 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-secret-volume\") pod \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\" (UID: \"d729e7c6-4664-47cb-b4bd-9892e1cf5aec\") " Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.834788 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-config-volume" (OuterVolumeSpecName: "config-volume") pod "d729e7c6-4664-47cb-b4bd-9892e1cf5aec" (UID: "d729e7c6-4664-47cb-b4bd-9892e1cf5aec"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.838192 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-kube-api-access-fql5q" (OuterVolumeSpecName: "kube-api-access-fql5q") pod "d729e7c6-4664-47cb-b4bd-9892e1cf5aec" (UID: "d729e7c6-4664-47cb-b4bd-9892e1cf5aec"). InnerVolumeSpecName "kube-api-access-fql5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.838745 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d729e7c6-4664-47cb-b4bd-9892e1cf5aec" (UID: "d729e7c6-4664-47cb-b4bd-9892e1cf5aec"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.936556 4874 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.936613 4874 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 12:45:03 crc kubenswrapper[4874]: I0122 12:45:03.936631 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fql5q\" (UniqueName: \"kubernetes.io/projected/d729e7c6-4664-47cb-b4bd-9892e1cf5aec-kube-api-access-fql5q\") on node \"crc\" DevicePath \"\"" Jan 22 12:45:04 crc kubenswrapper[4874]: I0122 12:45:04.354443 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" event={"ID":"d729e7c6-4664-47cb-b4bd-9892e1cf5aec","Type":"ContainerDied","Data":"909e7447c9e1e506cd07baba43a99e331ccc847b174d0ffcaffa649c2276d6a9"} Jan 22 12:45:04 crc kubenswrapper[4874]: I0122 12:45:04.354504 4874 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="909e7447c9e1e506cd07baba43a99e331ccc847b174d0ffcaffa649c2276d6a9" Jan 22 12:45:04 crc kubenswrapper[4874]: I0122 12:45:04.354513 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484765-kr6bl" Jan 22 12:45:04 crc kubenswrapper[4874]: I0122 12:45:04.430488 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2"] Jan 22 12:45:04 crc kubenswrapper[4874]: I0122 12:45:04.436876 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484720-8vrb2"] Jan 22 12:45:04 crc kubenswrapper[4874]: I0122 12:45:04.734063 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" path="/var/lib/kubelet/pods/402518c1-cf9f-4e51-9e46-2c7d1527ce6d/volumes" Jan 22 12:45:04 crc kubenswrapper[4874]: I0122 12:45:04.735155 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b40805b8-a40b-4854-80b3-28916f4f8a43" path="/var/lib/kubelet/pods/b40805b8-a40b-4854-80b3-28916f4f8a43/volumes" Jan 22 12:45:14 crc kubenswrapper[4874]: I0122 12:45:14.716250 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:45:14 crc kubenswrapper[4874]: E0122 12:45:14.716744 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:45:25 crc kubenswrapper[4874]: I0122 12:45:25.716339 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:45:25 crc kubenswrapper[4874]: E0122 12:45:25.717100 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:45:34 crc kubenswrapper[4874]: I0122 12:45:34.055272 4874 scope.go:117] "RemoveContainer" containerID="8a25148d438b262cdaed2dfa6ed49fd79b1992854289d94b46eaf3e9b7e53219" Jan 22 12:45:36 crc kubenswrapper[4874]: I0122 12:45:36.723504 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:45:36 crc kubenswrapper[4874]: E0122 12:45:36.724091 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:45:47 crc kubenswrapper[4874]: I0122 12:45:47.715532 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:45:47 crc kubenswrapper[4874]: E0122 12:45:47.716172 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:46:02 crc kubenswrapper[4874]: I0122 12:46:02.715929 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:46:02 crc kubenswrapper[4874]: E0122 12:46:02.716711 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:46:17 crc kubenswrapper[4874]: I0122 12:46:17.715843 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:46:18 crc kubenswrapper[4874]: I0122 12:46:18.023066 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"53dc1afb13dd727722ac5bb9850ed1be0aeb705ab0c93f80550ac8600fdf68e8"} Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.475133 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c2lw8"] Jan 22 12:48:24 crc kubenswrapper[4874]: E0122 12:48:24.475889 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d729e7c6-4664-47cb-b4bd-9892e1cf5aec" containerName="collect-profiles" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.475971 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="d729e7c6-4664-47cb-b4bd-9892e1cf5aec" containerName="collect-profiles" Jan 22 12:48:24 crc kubenswrapper[4874]: E0122 12:48:24.475997 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" containerName="extract-content" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.476006 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" containerName="extract-content" Jan 22 12:48:24 crc kubenswrapper[4874]: E0122 12:48:24.476024 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" containerName="registry-server" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.476032 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" containerName="registry-server" Jan 22 12:48:24 crc kubenswrapper[4874]: E0122 12:48:24.476042 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" containerName="extract-utilities" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.476050 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" containerName="extract-utilities" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.476200 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="d729e7c6-4664-47cb-b4bd-9892e1cf5aec" containerName="collect-profiles" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.476222 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="402518c1-cf9f-4e51-9e46-2c7d1527ce6d" containerName="registry-server" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.477307 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.501274 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c2lw8"] Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.575410 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz8jv\" (UniqueName: \"kubernetes.io/projected/df5f75a7-68cf-4efb-85d6-663e9e293c0d-kube-api-access-kz8jv\") pod \"community-operators-c2lw8\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.575508 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-utilities\") pod \"community-operators-c2lw8\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.575529 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-catalog-content\") pod \"community-operators-c2lw8\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.676899 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz8jv\" (UniqueName: \"kubernetes.io/projected/df5f75a7-68cf-4efb-85d6-663e9e293c0d-kube-api-access-kz8jv\") pod \"community-operators-c2lw8\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.677008 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-utilities\") pod \"community-operators-c2lw8\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.677031 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-catalog-content\") pod \"community-operators-c2lw8\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.677630 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-catalog-content\") pod \"community-operators-c2lw8\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.677647 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-utilities\") pod \"community-operators-c2lw8\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.710458 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz8jv\" (UniqueName: \"kubernetes.io/projected/df5f75a7-68cf-4efb-85d6-663e9e293c0d-kube-api-access-kz8jv\") pod \"community-operators-c2lw8\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:24 crc kubenswrapper[4874]: I0122 12:48:24.848776 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:25 crc kubenswrapper[4874]: I0122 12:48:25.154964 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c2lw8"] Jan 22 12:48:26 crc kubenswrapper[4874]: I0122 12:48:26.065734 4874 generic.go:334] "Generic (PLEG): container finished" podID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" containerID="169a02c51c136bfbf3fb31b59c7139a015d5b38327a437427f92cfbb15fd2587" exitCode=0 Jan 22 12:48:26 crc kubenswrapper[4874]: I0122 12:48:26.065809 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2lw8" event={"ID":"df5f75a7-68cf-4efb-85d6-663e9e293c0d","Type":"ContainerDied","Data":"169a02c51c136bfbf3fb31b59c7139a015d5b38327a437427f92cfbb15fd2587"} Jan 22 12:48:26 crc kubenswrapper[4874]: I0122 12:48:26.066170 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2lw8" event={"ID":"df5f75a7-68cf-4efb-85d6-663e9e293c0d","Type":"ContainerStarted","Data":"cb79405ab1fef9062efed21bf67b07217b663478707cfe8835b6a9c305df5370"} Jan 22 12:48:27 crc kubenswrapper[4874]: I0122 12:48:27.096679 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2lw8" event={"ID":"df5f75a7-68cf-4efb-85d6-663e9e293c0d","Type":"ContainerStarted","Data":"96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e"} Jan 22 12:48:28 crc kubenswrapper[4874]: I0122 12:48:28.106223 4874 generic.go:334] "Generic (PLEG): container finished" podID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" containerID="96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e" exitCode=0 Jan 22 12:48:28 crc kubenswrapper[4874]: I0122 12:48:28.106317 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2lw8" event={"ID":"df5f75a7-68cf-4efb-85d6-663e9e293c0d","Type":"ContainerDied","Data":"96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e"} Jan 22 12:48:28 crc kubenswrapper[4874]: I0122 12:48:28.106680 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2lw8" event={"ID":"df5f75a7-68cf-4efb-85d6-663e9e293c0d","Type":"ContainerStarted","Data":"7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b"} Jan 22 12:48:29 crc kubenswrapper[4874]: I0122 12:48:29.145885 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c2lw8" podStartSLOduration=3.462193242 podStartE2EDuration="5.145865566s" podCreationTimestamp="2026-01-22 12:48:24 +0000 UTC" firstStartedPulling="2026-01-22 12:48:26.067824691 +0000 UTC m=+4079.912895761" lastFinishedPulling="2026-01-22 12:48:27.751496995 +0000 UTC m=+4081.596568085" observedRunningTime="2026-01-22 12:48:29.139877931 +0000 UTC m=+4082.984949031" watchObservedRunningTime="2026-01-22 12:48:29.145865566 +0000 UTC m=+4082.990936656" Jan 22 12:48:34 crc kubenswrapper[4874]: I0122 12:48:34.849670 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:34 crc kubenswrapper[4874]: I0122 12:48:34.850182 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:34 crc kubenswrapper[4874]: I0122 12:48:34.914124 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:35 crc kubenswrapper[4874]: I0122 12:48:35.212498 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:35 crc kubenswrapper[4874]: I0122 12:48:35.275475 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c2lw8"] Jan 22 12:48:37 crc kubenswrapper[4874]: I0122 12:48:37.182933 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c2lw8" podUID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" containerName="registry-server" containerID="cri-o://7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b" gracePeriod=2 Jan 22 12:48:38 crc kubenswrapper[4874]: I0122 12:48:38.737496 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:38 crc kubenswrapper[4874]: I0122 12:48:38.918914 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-catalog-content\") pod \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " Jan 22 12:48:38 crc kubenswrapper[4874]: I0122 12:48:38.919021 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz8jv\" (UniqueName: \"kubernetes.io/projected/df5f75a7-68cf-4efb-85d6-663e9e293c0d-kube-api-access-kz8jv\") pod \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " Jan 22 12:48:38 crc kubenswrapper[4874]: I0122 12:48:38.919060 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-utilities\") pod \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\" (UID: \"df5f75a7-68cf-4efb-85d6-663e9e293c0d\") " Jan 22 12:48:38 crc kubenswrapper[4874]: I0122 12:48:38.920284 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-utilities" (OuterVolumeSpecName: "utilities") pod "df5f75a7-68cf-4efb-85d6-663e9e293c0d" (UID: "df5f75a7-68cf-4efb-85d6-663e9e293c0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:48:38 crc kubenswrapper[4874]: I0122 12:48:38.930419 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df5f75a7-68cf-4efb-85d6-663e9e293c0d-kube-api-access-kz8jv" (OuterVolumeSpecName: "kube-api-access-kz8jv") pod "df5f75a7-68cf-4efb-85d6-663e9e293c0d" (UID: "df5f75a7-68cf-4efb-85d6-663e9e293c0d"). InnerVolumeSpecName "kube-api-access-kz8jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:48:38 crc kubenswrapper[4874]: I0122 12:48:38.995754 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df5f75a7-68cf-4efb-85d6-663e9e293c0d" (UID: "df5f75a7-68cf-4efb-85d6-663e9e293c0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.021055 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz8jv\" (UniqueName: \"kubernetes.io/projected/df5f75a7-68cf-4efb-85d6-663e9e293c0d-kube-api-access-kz8jv\") on node \"crc\" DevicePath \"\"" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.021093 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.021104 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5f75a7-68cf-4efb-85d6-663e9e293c0d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.205774 4874 generic.go:334] "Generic (PLEG): container finished" podID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" containerID="7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b" exitCode=0 Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.205824 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2lw8" event={"ID":"df5f75a7-68cf-4efb-85d6-663e9e293c0d","Type":"ContainerDied","Data":"7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b"} Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.205853 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2lw8" event={"ID":"df5f75a7-68cf-4efb-85d6-663e9e293c0d","Type":"ContainerDied","Data":"cb79405ab1fef9062efed21bf67b07217b663478707cfe8835b6a9c305df5370"} Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.205873 4874 scope.go:117] "RemoveContainer" containerID="7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.206013 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c2lw8" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.266037 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c2lw8"] Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.273377 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c2lw8"] Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.276389 4874 scope.go:117] "RemoveContainer" containerID="96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.297369 4874 scope.go:117] "RemoveContainer" containerID="169a02c51c136bfbf3fb31b59c7139a015d5b38327a437427f92cfbb15fd2587" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.325549 4874 scope.go:117] "RemoveContainer" containerID="7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b" Jan 22 12:48:39 crc kubenswrapper[4874]: E0122 12:48:39.326161 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b\": container with ID starting with 7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b not found: ID does not exist" containerID="7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.326223 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b"} err="failed to get container status \"7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b\": rpc error: code = NotFound desc = could not find container \"7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b\": container with ID starting with 7c4043ead839e73ba762859fea2607e4f3556e5ca613450f714358fcc2dabd7b not found: ID does not exist" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.326250 4874 scope.go:117] "RemoveContainer" containerID="96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e" Jan 22 12:48:39 crc kubenswrapper[4874]: E0122 12:48:39.326738 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e\": container with ID starting with 96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e not found: ID does not exist" containerID="96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.326772 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e"} err="failed to get container status \"96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e\": rpc error: code = NotFound desc = could not find container \"96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e\": container with ID starting with 96f19a91649c00d3c5d916c4724cdf97877eaf4c77feceb978fd362291601a3e not found: ID does not exist" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.326793 4874 scope.go:117] "RemoveContainer" containerID="169a02c51c136bfbf3fb31b59c7139a015d5b38327a437427f92cfbb15fd2587" Jan 22 12:48:39 crc kubenswrapper[4874]: E0122 12:48:39.327268 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"169a02c51c136bfbf3fb31b59c7139a015d5b38327a437427f92cfbb15fd2587\": container with ID starting with 169a02c51c136bfbf3fb31b59c7139a015d5b38327a437427f92cfbb15fd2587 not found: ID does not exist" containerID="169a02c51c136bfbf3fb31b59c7139a015d5b38327a437427f92cfbb15fd2587" Jan 22 12:48:39 crc kubenswrapper[4874]: I0122 12:48:39.327293 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"169a02c51c136bfbf3fb31b59c7139a015d5b38327a437427f92cfbb15fd2587"} err="failed to get container status \"169a02c51c136bfbf3fb31b59c7139a015d5b38327a437427f92cfbb15fd2587\": rpc error: code = NotFound desc = could not find container \"169a02c51c136bfbf3fb31b59c7139a015d5b38327a437427f92cfbb15fd2587\": container with ID starting with 169a02c51c136bfbf3fb31b59c7139a015d5b38327a437427f92cfbb15fd2587 not found: ID does not exist" Jan 22 12:48:40 crc kubenswrapper[4874]: I0122 12:48:40.729179 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" path="/var/lib/kubelet/pods/df5f75a7-68cf-4efb-85d6-663e9e293c0d/volumes" Jan 22 12:48:43 crc kubenswrapper[4874]: I0122 12:48:43.520885 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:48:43 crc kubenswrapper[4874]: I0122 12:48:43.521241 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:49:13 crc kubenswrapper[4874]: I0122 12:49:13.520811 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:49:13 crc kubenswrapper[4874]: I0122 12:49:13.521598 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:49:43 crc kubenswrapper[4874]: I0122 12:49:43.520566 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:49:43 crc kubenswrapper[4874]: I0122 12:49:43.521009 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:49:43 crc kubenswrapper[4874]: I0122 12:49:43.521052 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 12:49:43 crc kubenswrapper[4874]: I0122 12:49:43.521667 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"53dc1afb13dd727722ac5bb9850ed1be0aeb705ab0c93f80550ac8600fdf68e8"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 12:49:43 crc kubenswrapper[4874]: I0122 12:49:43.521713 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://53dc1afb13dd727722ac5bb9850ed1be0aeb705ab0c93f80550ac8600fdf68e8" gracePeriod=600 Jan 22 12:49:43 crc kubenswrapper[4874]: I0122 12:49:43.810206 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="53dc1afb13dd727722ac5bb9850ed1be0aeb705ab0c93f80550ac8600fdf68e8" exitCode=0 Jan 22 12:49:43 crc kubenswrapper[4874]: I0122 12:49:43.810266 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"53dc1afb13dd727722ac5bb9850ed1be0aeb705ab0c93f80550ac8600fdf68e8"} Jan 22 12:49:43 crc kubenswrapper[4874]: I0122 12:49:43.810641 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerStarted","Data":"512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96"} Jan 22 12:49:43 crc kubenswrapper[4874]: I0122 12:49:43.810666 4874 scope.go:117] "RemoveContainer" containerID="f8b0481d5958c4e3126d530e9c4112219efd18dd96b27a8af7b05bee45a4994b" Jan 22 12:51:43 crc kubenswrapper[4874]: I0122 12:51:43.521117 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:51:43 crc kubenswrapper[4874]: I0122 12:51:43.521889 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:52:13 crc kubenswrapper[4874]: I0122 12:52:13.520888 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:52:13 crc kubenswrapper[4874]: I0122 12:52:13.521465 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.427162 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8j5sk"] Jan 22 12:52:15 crc kubenswrapper[4874]: E0122 12:52:15.427706 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" containerName="extract-content" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.427735 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" containerName="extract-content" Jan 22 12:52:15 crc kubenswrapper[4874]: E0122 12:52:15.427775 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" containerName="extract-utilities" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.427792 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" containerName="extract-utilities" Jan 22 12:52:15 crc kubenswrapper[4874]: E0122 12:52:15.427822 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" containerName="registry-server" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.427838 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" containerName="registry-server" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.428070 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="df5f75a7-68cf-4efb-85d6-663e9e293c0d" containerName="registry-server" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.429770 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.450255 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8j5sk"] Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.457547 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-utilities\") pod \"redhat-operators-8j5sk\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.457665 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-catalog-content\") pod \"redhat-operators-8j5sk\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.457758 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m9ft\" (UniqueName: \"kubernetes.io/projected/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-kube-api-access-6m9ft\") pod \"redhat-operators-8j5sk\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.558790 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-utilities\") pod \"redhat-operators-8j5sk\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.558838 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-catalog-content\") pod \"redhat-operators-8j5sk\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.558873 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m9ft\" (UniqueName: \"kubernetes.io/projected/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-kube-api-access-6m9ft\") pod \"redhat-operators-8j5sk\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.559470 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-utilities\") pod \"redhat-operators-8j5sk\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.559499 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-catalog-content\") pod \"redhat-operators-8j5sk\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.585011 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m9ft\" (UniqueName: \"kubernetes.io/projected/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-kube-api-access-6m9ft\") pod \"redhat-operators-8j5sk\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.761965 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:15 crc kubenswrapper[4874]: I0122 12:52:15.982167 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8j5sk"] Jan 22 12:52:16 crc kubenswrapper[4874]: I0122 12:52:16.030341 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8j5sk" event={"ID":"7aa4b82e-4df8-48e6-a5a7-7994665e26c7","Type":"ContainerStarted","Data":"07f74157b807b2a05eba4b8e26ed2460aa0f21a6ccc677b52bf5605d386ba910"} Jan 22 12:52:17 crc kubenswrapper[4874]: I0122 12:52:17.037877 4874 generic.go:334] "Generic (PLEG): container finished" podID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerID="eb2a19caecc5454a93f49fc4f9ab75a3fe53e6bf338ab8ae7b3522cca4fb0fc6" exitCode=0 Jan 22 12:52:17 crc kubenswrapper[4874]: I0122 12:52:17.038133 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8j5sk" event={"ID":"7aa4b82e-4df8-48e6-a5a7-7994665e26c7","Type":"ContainerDied","Data":"eb2a19caecc5454a93f49fc4f9ab75a3fe53e6bf338ab8ae7b3522cca4fb0fc6"} Jan 22 12:52:17 crc kubenswrapper[4874]: I0122 12:52:17.040317 4874 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 12:52:19 crc kubenswrapper[4874]: I0122 12:52:19.057316 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8j5sk" event={"ID":"7aa4b82e-4df8-48e6-a5a7-7994665e26c7","Type":"ContainerStarted","Data":"a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2"} Jan 22 12:52:22 crc kubenswrapper[4874]: I0122 12:52:22.079563 4874 generic.go:334] "Generic (PLEG): container finished" podID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerID="a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2" exitCode=0 Jan 22 12:52:22 crc kubenswrapper[4874]: I0122 12:52:22.079638 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8j5sk" event={"ID":"7aa4b82e-4df8-48e6-a5a7-7994665e26c7","Type":"ContainerDied","Data":"a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2"} Jan 22 12:52:24 crc kubenswrapper[4874]: I0122 12:52:24.095416 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8j5sk" event={"ID":"7aa4b82e-4df8-48e6-a5a7-7994665e26c7","Type":"ContainerStarted","Data":"b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147"} Jan 22 12:52:24 crc kubenswrapper[4874]: I0122 12:52:24.125117 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8j5sk" podStartSLOduration=3.360829217 podStartE2EDuration="9.125103145s" podCreationTimestamp="2026-01-22 12:52:15 +0000 UTC" firstStartedPulling="2026-01-22 12:52:17.040038918 +0000 UTC m=+4310.885109988" lastFinishedPulling="2026-01-22 12:52:22.804312836 +0000 UTC m=+4316.649383916" observedRunningTime="2026-01-22 12:52:24.121032879 +0000 UTC m=+4317.966103949" watchObservedRunningTime="2026-01-22 12:52:24.125103145 +0000 UTC m=+4317.970174215" Jan 22 12:52:25 crc kubenswrapper[4874]: I0122 12:52:25.763507 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:25 crc kubenswrapper[4874]: I0122 12:52:25.763766 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:26 crc kubenswrapper[4874]: I0122 12:52:26.818351 4874 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8j5sk" podUID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerName="registry-server" probeResult="failure" output=< Jan 22 12:52:26 crc kubenswrapper[4874]: timeout: failed to connect service ":50051" within 1s Jan 22 12:52:26 crc kubenswrapper[4874]: > Jan 22 12:52:35 crc kubenswrapper[4874]: I0122 12:52:35.802259 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:35 crc kubenswrapper[4874]: I0122 12:52:35.863282 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:36 crc kubenswrapper[4874]: I0122 12:52:36.034136 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8j5sk"] Jan 22 12:52:37 crc kubenswrapper[4874]: I0122 12:52:37.193354 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8j5sk" podUID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerName="registry-server" containerID="cri-o://b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147" gracePeriod=2 Jan 22 12:52:37 crc kubenswrapper[4874]: I0122 12:52:37.582731 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:37 crc kubenswrapper[4874]: I0122 12:52:37.675915 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-utilities\") pod \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " Jan 22 12:52:37 crc kubenswrapper[4874]: I0122 12:52:37.675992 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m9ft\" (UniqueName: \"kubernetes.io/projected/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-kube-api-access-6m9ft\") pod \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " Jan 22 12:52:37 crc kubenswrapper[4874]: I0122 12:52:37.677047 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-utilities" (OuterVolumeSpecName: "utilities") pod "7aa4b82e-4df8-48e6-a5a7-7994665e26c7" (UID: "7aa4b82e-4df8-48e6-a5a7-7994665e26c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:52:37 crc kubenswrapper[4874]: I0122 12:52:37.677287 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-catalog-content\") pod \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\" (UID: \"7aa4b82e-4df8-48e6-a5a7-7994665e26c7\") " Jan 22 12:52:37 crc kubenswrapper[4874]: I0122 12:52:37.677679 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:52:37 crc kubenswrapper[4874]: I0122 12:52:37.681462 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-kube-api-access-6m9ft" (OuterVolumeSpecName: "kube-api-access-6m9ft") pod "7aa4b82e-4df8-48e6-a5a7-7994665e26c7" (UID: "7aa4b82e-4df8-48e6-a5a7-7994665e26c7"). InnerVolumeSpecName "kube-api-access-6m9ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:52:37 crc kubenswrapper[4874]: I0122 12:52:37.778690 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m9ft\" (UniqueName: \"kubernetes.io/projected/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-kube-api-access-6m9ft\") on node \"crc\" DevicePath \"\"" Jan 22 12:52:37 crc kubenswrapper[4874]: I0122 12:52:37.802349 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7aa4b82e-4df8-48e6-a5a7-7994665e26c7" (UID: "7aa4b82e-4df8-48e6-a5a7-7994665e26c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:52:37 crc kubenswrapper[4874]: I0122 12:52:37.880518 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7aa4b82e-4df8-48e6-a5a7-7994665e26c7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.219706 4874 generic.go:334] "Generic (PLEG): container finished" podID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerID="b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147" exitCode=0 Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.219795 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8j5sk" event={"ID":"7aa4b82e-4df8-48e6-a5a7-7994665e26c7","Type":"ContainerDied","Data":"b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147"} Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.219854 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8j5sk" Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.221285 4874 scope.go:117] "RemoveContainer" containerID="b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147" Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.228456 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8j5sk" event={"ID":"7aa4b82e-4df8-48e6-a5a7-7994665e26c7","Type":"ContainerDied","Data":"07f74157b807b2a05eba4b8e26ed2460aa0f21a6ccc677b52bf5605d386ba910"} Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.242545 4874 scope.go:117] "RemoveContainer" containerID="a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2" Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.262610 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8j5sk"] Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.268140 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8j5sk"] Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.287940 4874 scope.go:117] "RemoveContainer" containerID="eb2a19caecc5454a93f49fc4f9ab75a3fe53e6bf338ab8ae7b3522cca4fb0fc6" Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.303492 4874 scope.go:117] "RemoveContainer" containerID="b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147" Jan 22 12:52:38 crc kubenswrapper[4874]: E0122 12:52:38.303952 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147\": container with ID starting with b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147 not found: ID does not exist" containerID="b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147" Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.303989 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147"} err="failed to get container status \"b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147\": rpc error: code = NotFound desc = could not find container \"b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147\": container with ID starting with b65033740bb3015260b8a13295eef496a81ccc449e06c0bd433357bda1460147 not found: ID does not exist" Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.304018 4874 scope.go:117] "RemoveContainer" containerID="a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2" Jan 22 12:52:38 crc kubenswrapper[4874]: E0122 12:52:38.304265 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2\": container with ID starting with a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2 not found: ID does not exist" containerID="a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2" Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.304290 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2"} err="failed to get container status \"a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2\": rpc error: code = NotFound desc = could not find container \"a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2\": container with ID starting with a2187d719b1797b03f7c8469e92ae17b306634dd3fd22fabb35efaaa87135ae2 not found: ID does not exist" Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.304308 4874 scope.go:117] "RemoveContainer" containerID="eb2a19caecc5454a93f49fc4f9ab75a3fe53e6bf338ab8ae7b3522cca4fb0fc6" Jan 22 12:52:38 crc kubenswrapper[4874]: E0122 12:52:38.304578 4874 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb2a19caecc5454a93f49fc4f9ab75a3fe53e6bf338ab8ae7b3522cca4fb0fc6\": container with ID starting with eb2a19caecc5454a93f49fc4f9ab75a3fe53e6bf338ab8ae7b3522cca4fb0fc6 not found: ID does not exist" containerID="eb2a19caecc5454a93f49fc4f9ab75a3fe53e6bf338ab8ae7b3522cca4fb0fc6" Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.304602 4874 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb2a19caecc5454a93f49fc4f9ab75a3fe53e6bf338ab8ae7b3522cca4fb0fc6"} err="failed to get container status \"eb2a19caecc5454a93f49fc4f9ab75a3fe53e6bf338ab8ae7b3522cca4fb0fc6\": rpc error: code = NotFound desc = could not find container \"eb2a19caecc5454a93f49fc4f9ab75a3fe53e6bf338ab8ae7b3522cca4fb0fc6\": container with ID starting with eb2a19caecc5454a93f49fc4f9ab75a3fe53e6bf338ab8ae7b3522cca4fb0fc6 not found: ID does not exist" Jan 22 12:52:38 crc kubenswrapper[4874]: I0122 12:52:38.726120 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" path="/var/lib/kubelet/pods/7aa4b82e-4df8-48e6-a5a7-7994665e26c7/volumes" Jan 22 12:52:43 crc kubenswrapper[4874]: I0122 12:52:43.520994 4874 patch_prober.go:28] interesting pod/machine-config-daemon-4prkg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 12:52:43 crc kubenswrapper[4874]: I0122 12:52:43.521696 4874 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 12:52:43 crc kubenswrapper[4874]: I0122 12:52:43.521753 4874 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" Jan 22 12:52:43 crc kubenswrapper[4874]: I0122 12:52:43.522381 4874 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96"} pod="openshift-machine-config-operator/machine-config-daemon-4prkg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 12:52:43 crc kubenswrapper[4874]: I0122 12:52:43.522445 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerName="machine-config-daemon" containerID="cri-o://512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" gracePeriod=600 Jan 22 12:52:43 crc kubenswrapper[4874]: E0122 12:52:43.653192 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:52:44 crc kubenswrapper[4874]: I0122 12:52:44.274328 4874 generic.go:334] "Generic (PLEG): container finished" podID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" exitCode=0 Jan 22 12:52:44 crc kubenswrapper[4874]: I0122 12:52:44.274737 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" event={"ID":"7c9653f9-cd5b-4b7a-8056-80ae8235d039","Type":"ContainerDied","Data":"512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96"} Jan 22 12:52:44 crc kubenswrapper[4874]: I0122 12:52:44.274794 4874 scope.go:117] "RemoveContainer" containerID="53dc1afb13dd727722ac5bb9850ed1be0aeb705ab0c93f80550ac8600fdf68e8" Jan 22 12:52:44 crc kubenswrapper[4874]: I0122 12:52:44.275389 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:52:44 crc kubenswrapper[4874]: E0122 12:52:44.275697 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:52:58 crc kubenswrapper[4874]: I0122 12:52:58.716643 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:52:58 crc kubenswrapper[4874]: E0122 12:52:58.717231 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:53:09 crc kubenswrapper[4874]: I0122 12:53:09.716851 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:53:09 crc kubenswrapper[4874]: E0122 12:53:09.718059 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:53:23 crc kubenswrapper[4874]: I0122 12:53:23.716095 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:53:23 crc kubenswrapper[4874]: E0122 12:53:23.717064 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:53:34 crc kubenswrapper[4874]: I0122 12:53:34.718955 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:53:34 crc kubenswrapper[4874]: E0122 12:53:34.719827 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:53:48 crc kubenswrapper[4874]: I0122 12:53:48.716002 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:53:48 crc kubenswrapper[4874]: E0122 12:53:48.716796 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:53:59 crc kubenswrapper[4874]: I0122 12:53:59.716568 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:53:59 crc kubenswrapper[4874]: E0122 12:53:59.717311 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:54:14 crc kubenswrapper[4874]: I0122 12:54:14.716876 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:54:14 crc kubenswrapper[4874]: E0122 12:54:14.718126 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:54:25 crc kubenswrapper[4874]: I0122 12:54:25.716281 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:54:25 crc kubenswrapper[4874]: E0122 12:54:25.717005 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:54:37 crc kubenswrapper[4874]: I0122 12:54:37.716562 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:54:37 crc kubenswrapper[4874]: E0122 12:54:37.717282 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.712202 4874 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4ts5j"] Jan 22 12:54:49 crc kubenswrapper[4874]: E0122 12:54:49.713030 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerName="extract-content" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.713044 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerName="extract-content" Jan 22 12:54:49 crc kubenswrapper[4874]: E0122 12:54:49.713062 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerName="registry-server" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.713068 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerName="registry-server" Jan 22 12:54:49 crc kubenswrapper[4874]: E0122 12:54:49.713086 4874 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerName="extract-utilities" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.713092 4874 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerName="extract-utilities" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.713224 4874 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aa4b82e-4df8-48e6-a5a7-7994665e26c7" containerName="registry-server" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.714311 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.731445 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4ts5j"] Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.806465 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nr2n\" (UniqueName: \"kubernetes.io/projected/3366cea5-606e-469e-8088-f7746fcd4ac8-kube-api-access-8nr2n\") pod \"certified-operators-4ts5j\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.806529 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-utilities\") pod \"certified-operators-4ts5j\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.806732 4874 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-catalog-content\") pod \"certified-operators-4ts5j\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.908322 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-catalog-content\") pod \"certified-operators-4ts5j\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.908426 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nr2n\" (UniqueName: \"kubernetes.io/projected/3366cea5-606e-469e-8088-f7746fcd4ac8-kube-api-access-8nr2n\") pod \"certified-operators-4ts5j\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.908463 4874 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-utilities\") pod \"certified-operators-4ts5j\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.908980 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-catalog-content\") pod \"certified-operators-4ts5j\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.909014 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-utilities\") pod \"certified-operators-4ts5j\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:54:49 crc kubenswrapper[4874]: I0122 12:54:49.929022 4874 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nr2n\" (UniqueName: \"kubernetes.io/projected/3366cea5-606e-469e-8088-f7746fcd4ac8-kube-api-access-8nr2n\") pod \"certified-operators-4ts5j\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:54:50 crc kubenswrapper[4874]: I0122 12:54:50.032253 4874 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:54:50 crc kubenswrapper[4874]: I0122 12:54:50.313381 4874 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4ts5j"] Jan 22 12:54:51 crc kubenswrapper[4874]: I0122 12:54:51.300186 4874 generic.go:334] "Generic (PLEG): container finished" podID="3366cea5-606e-469e-8088-f7746fcd4ac8" containerID="db1f2c79c07d02974d8a2cb5c4f9b17ce3f67cd346cb4e030c7d8b038100e7f7" exitCode=0 Jan 22 12:54:51 crc kubenswrapper[4874]: I0122 12:54:51.300305 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ts5j" event={"ID":"3366cea5-606e-469e-8088-f7746fcd4ac8","Type":"ContainerDied","Data":"db1f2c79c07d02974d8a2cb5c4f9b17ce3f67cd346cb4e030c7d8b038100e7f7"} Jan 22 12:54:51 crc kubenswrapper[4874]: I0122 12:54:51.300782 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ts5j" event={"ID":"3366cea5-606e-469e-8088-f7746fcd4ac8","Type":"ContainerStarted","Data":"41cc6f8e5bcc4e68b001c69fbdccd567601b6df04c427c9ee944bd141d5b9949"} Jan 22 12:54:51 crc kubenswrapper[4874]: I0122 12:54:51.717007 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:54:51 crc kubenswrapper[4874]: E0122 12:54:51.717180 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:54:52 crc kubenswrapper[4874]: I0122 12:54:52.311497 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ts5j" event={"ID":"3366cea5-606e-469e-8088-f7746fcd4ac8","Type":"ContainerStarted","Data":"fdfeab215538aa7f0b4981c2530a33659540800849db1e24078e5d1519bc776a"} Jan 22 12:54:53 crc kubenswrapper[4874]: I0122 12:54:53.320488 4874 generic.go:334] "Generic (PLEG): container finished" podID="3366cea5-606e-469e-8088-f7746fcd4ac8" containerID="fdfeab215538aa7f0b4981c2530a33659540800849db1e24078e5d1519bc776a" exitCode=0 Jan 22 12:54:53 crc kubenswrapper[4874]: I0122 12:54:53.320809 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ts5j" event={"ID":"3366cea5-606e-469e-8088-f7746fcd4ac8","Type":"ContainerDied","Data":"fdfeab215538aa7f0b4981c2530a33659540800849db1e24078e5d1519bc776a"} Jan 22 12:54:54 crc kubenswrapper[4874]: I0122 12:54:54.329502 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ts5j" event={"ID":"3366cea5-606e-469e-8088-f7746fcd4ac8","Type":"ContainerStarted","Data":"f7005ebc37b8caea79a0a57a81efc611604529a73f8eabae9037f86dc40b1ccf"} Jan 22 12:54:54 crc kubenswrapper[4874]: I0122 12:54:54.348956 4874 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4ts5j" podStartSLOduration=2.756714098 podStartE2EDuration="5.348935565s" podCreationTimestamp="2026-01-22 12:54:49 +0000 UTC" firstStartedPulling="2026-01-22 12:54:51.304208942 +0000 UTC m=+4465.149280022" lastFinishedPulling="2026-01-22 12:54:53.896430419 +0000 UTC m=+4467.741501489" observedRunningTime="2026-01-22 12:54:54.344122547 +0000 UTC m=+4468.189193617" watchObservedRunningTime="2026-01-22 12:54:54.348935565 +0000 UTC m=+4468.194006635" Jan 22 12:55:00 crc kubenswrapper[4874]: I0122 12:55:00.032949 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:55:00 crc kubenswrapper[4874]: I0122 12:55:00.033702 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:55:00 crc kubenswrapper[4874]: I0122 12:55:00.088592 4874 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:55:00 crc kubenswrapper[4874]: I0122 12:55:00.433828 4874 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:55:00 crc kubenswrapper[4874]: I0122 12:55:00.483484 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4ts5j"] Jan 22 12:55:02 crc kubenswrapper[4874]: I0122 12:55:02.395911 4874 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4ts5j" podUID="3366cea5-606e-469e-8088-f7746fcd4ac8" containerName="registry-server" containerID="cri-o://f7005ebc37b8caea79a0a57a81efc611604529a73f8eabae9037f86dc40b1ccf" gracePeriod=2 Jan 22 12:55:03 crc kubenswrapper[4874]: I0122 12:55:03.715513 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:55:03 crc kubenswrapper[4874]: E0122 12:55:03.715751 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.418333 4874 generic.go:334] "Generic (PLEG): container finished" podID="3366cea5-606e-469e-8088-f7746fcd4ac8" containerID="f7005ebc37b8caea79a0a57a81efc611604529a73f8eabae9037f86dc40b1ccf" exitCode=0 Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.418427 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ts5j" event={"ID":"3366cea5-606e-469e-8088-f7746fcd4ac8","Type":"ContainerDied","Data":"f7005ebc37b8caea79a0a57a81efc611604529a73f8eabae9037f86dc40b1ccf"} Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.496972 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.645139 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-utilities\") pod \"3366cea5-606e-469e-8088-f7746fcd4ac8\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.645682 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-catalog-content\") pod \"3366cea5-606e-469e-8088-f7746fcd4ac8\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.645891 4874 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nr2n\" (UniqueName: \"kubernetes.io/projected/3366cea5-606e-469e-8088-f7746fcd4ac8-kube-api-access-8nr2n\") pod \"3366cea5-606e-469e-8088-f7746fcd4ac8\" (UID: \"3366cea5-606e-469e-8088-f7746fcd4ac8\") " Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.646083 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-utilities" (OuterVolumeSpecName: "utilities") pod "3366cea5-606e-469e-8088-f7746fcd4ac8" (UID: "3366cea5-606e-469e-8088-f7746fcd4ac8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.646765 4874 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.652858 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3366cea5-606e-469e-8088-f7746fcd4ac8-kube-api-access-8nr2n" (OuterVolumeSpecName: "kube-api-access-8nr2n") pod "3366cea5-606e-469e-8088-f7746fcd4ac8" (UID: "3366cea5-606e-469e-8088-f7746fcd4ac8"). InnerVolumeSpecName "kube-api-access-8nr2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.692113 4874 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3366cea5-606e-469e-8088-f7746fcd4ac8" (UID: "3366cea5-606e-469e-8088-f7746fcd4ac8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.747796 4874 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3366cea5-606e-469e-8088-f7746fcd4ac8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 12:55:05 crc kubenswrapper[4874]: I0122 12:55:05.747830 4874 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nr2n\" (UniqueName: \"kubernetes.io/projected/3366cea5-606e-469e-8088-f7746fcd4ac8-kube-api-access-8nr2n\") on node \"crc\" DevicePath \"\"" Jan 22 12:55:06 crc kubenswrapper[4874]: I0122 12:55:06.448946 4874 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ts5j" event={"ID":"3366cea5-606e-469e-8088-f7746fcd4ac8","Type":"ContainerDied","Data":"41cc6f8e5bcc4e68b001c69fbdccd567601b6df04c427c9ee944bd141d5b9949"} Jan 22 12:55:06 crc kubenswrapper[4874]: I0122 12:55:06.449030 4874 scope.go:117] "RemoveContainer" containerID="f7005ebc37b8caea79a0a57a81efc611604529a73f8eabae9037f86dc40b1ccf" Jan 22 12:55:06 crc kubenswrapper[4874]: I0122 12:55:06.449069 4874 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ts5j" Jan 22 12:55:06 crc kubenswrapper[4874]: I0122 12:55:06.481576 4874 scope.go:117] "RemoveContainer" containerID="fdfeab215538aa7f0b4981c2530a33659540800849db1e24078e5d1519bc776a" Jan 22 12:55:06 crc kubenswrapper[4874]: I0122 12:55:06.489509 4874 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4ts5j"] Jan 22 12:55:06 crc kubenswrapper[4874]: I0122 12:55:06.494992 4874 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4ts5j"] Jan 22 12:55:06 crc kubenswrapper[4874]: I0122 12:55:06.508149 4874 scope.go:117] "RemoveContainer" containerID="db1f2c79c07d02974d8a2cb5c4f9b17ce3f67cd346cb4e030c7d8b038100e7f7" Jan 22 12:55:06 crc kubenswrapper[4874]: I0122 12:55:06.729694 4874 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3366cea5-606e-469e-8088-f7746fcd4ac8" path="/var/lib/kubelet/pods/3366cea5-606e-469e-8088-f7746fcd4ac8/volumes" Jan 22 12:55:18 crc kubenswrapper[4874]: I0122 12:55:18.721297 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:55:18 crc kubenswrapper[4874]: E0122 12:55:18.722340 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" Jan 22 12:55:31 crc kubenswrapper[4874]: I0122 12:55:31.716176 4874 scope.go:117] "RemoveContainer" containerID="512f18641516777047c81cf22d0f34b66336e79fdf481778071e86dc745dae96" Jan 22 12:55:31 crc kubenswrapper[4874]: E0122 12:55:31.717175 4874 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4prkg_openshift-machine-config-operator(7c9653f9-cd5b-4b7a-8056-80ae8235d039)\"" pod="openshift-machine-config-operator/machine-config-daemon-4prkg" podUID="7c9653f9-cd5b-4b7a-8056-80ae8235d039" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515134417121024444 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015134417122017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015134405716016513 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015134405717015464 5ustar corecore